[ 470.175166] env[62235]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62235) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.175454] env[62235]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62235) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.175631] env[62235]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62235) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.175961] env[62235]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 470.268870] env[62235]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62235) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 470.278902] env[62235]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62235) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 470.878940] env[62235]: INFO nova.virt.driver [None req-64ae8919-df32-46b5-bc64-99c2a47ef092 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 470.951823] env[62235]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.951985] env[62235]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.952123] env[62235]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62235) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 474.017221] env[62235]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-5fbf2d78-69bd-444d-a455-8e11ffc1f4d9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.033935] env[62235]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62235) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 474.034157] env[62235]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-ebcf046c-c52b-4729-b2b9-38af1a46470a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.063240] env[62235]: INFO oslo_vmware.api [-] Successfully established new session; session ID is ebfda. [ 474.063384] env[62235]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.111s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.063923] env[62235]: INFO nova.virt.vmwareapi.driver [None req-64ae8919-df32-46b5-bc64-99c2a47ef092 None None] VMware vCenter version: 7.0.3 [ 474.067242] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abfd4ac-cbfa-4e75-8781-ddd4a4e9a45b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.084238] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde8528c-d190-40cd-8b94-3f2e8e5225aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.090041] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376d33ae-6c5e-4bd5-9872-017a07deb0bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.096630] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0e5d01-a55f-46db-913a-00425cfbc23a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.109629] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb37ee65-6766-42f9-8461-3acdd6b6a455 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.115612] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42661fd3-5833-4895-b0a8-59345b86dc0e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.146231] env[62235]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-defca50f-6c27-4362-84bd-803a75c2865b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.151450] env[62235]: DEBUG nova.virt.vmwareapi.driver [None req-64ae8919-df32-46b5-bc64-99c2a47ef092 None None] Extension org.openstack.compute already exists. {{(pid=62235) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 474.154344] env[62235]: INFO nova.compute.provider_config [None req-64ae8919-df32-46b5-bc64-99c2a47ef092 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 474.658008] env[62235]: DEBUG nova.context [None req-64ae8919-df32-46b5-bc64-99c2a47ef092 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),0c29467c-7d6f-4db3-a3ac-b809f3a59da0(cell1) {{(pid=62235) load_cells /opt/stack/nova/nova/context.py:464}} [ 474.660141] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.660379] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.661178] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.661658] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Acquiring lock "0c29467c-7d6f-4db3-a3ac-b809f3a59da0" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.661852] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Lock "0c29467c-7d6f-4db3-a3ac-b809f3a59da0" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.662871] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Lock "0c29467c-7d6f-4db3-a3ac-b809f3a59da0" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.682918] env[62235]: INFO dbcounter [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Registered counter for database nova_cell0 [ 474.691167] env[62235]: INFO dbcounter [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Registered counter for database nova_cell1 [ 474.694411] env[62235]: DEBUG oslo_db.sqlalchemy.engines [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62235) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 474.694754] env[62235]: DEBUG oslo_db.sqlalchemy.engines [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62235) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 474.699426] env[62235]: ERROR nova.db.main.api [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 474.699426] env[62235]: result = function(*args, **kwargs) [ 474.699426] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 474.699426] env[62235]: return func(*args, **kwargs) [ 474.699426] env[62235]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 474.699426] env[62235]: result = fn(*args, **kwargs) [ 474.699426] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 474.699426] env[62235]: return f(*args, **kwargs) [ 474.699426] env[62235]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 474.699426] env[62235]: return db.service_get_minimum_version(context, binaries) [ 474.699426] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 474.699426] env[62235]: _check_db_access() [ 474.699426] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 474.699426] env[62235]: stacktrace = ''.join(traceback.format_stack()) [ 474.699426] env[62235]: [ 474.700434] env[62235]: ERROR nova.db.main.api [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 474.700434] env[62235]: result = function(*args, **kwargs) [ 474.700434] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 474.700434] env[62235]: return func(*args, **kwargs) [ 474.700434] env[62235]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 474.700434] env[62235]: result = fn(*args, **kwargs) [ 474.700434] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 474.700434] env[62235]: return f(*args, **kwargs) [ 474.700434] env[62235]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 474.700434] env[62235]: return db.service_get_minimum_version(context, binaries) [ 474.700434] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 474.700434] env[62235]: _check_db_access() [ 474.700434] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 474.700434] env[62235]: stacktrace = ''.join(traceback.format_stack()) [ 474.700434] env[62235]: [ 474.700866] env[62235]: WARNING nova.objects.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Failed to get minimum service version for cell 0c29467c-7d6f-4db3-a3ac-b809f3a59da0 [ 474.700997] env[62235]: WARNING nova.objects.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 474.701390] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Acquiring lock "singleton_lock" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 474.701551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Acquired lock "singleton_lock" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 474.701800] env[62235]: DEBUG oslo_concurrency.lockutils [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Releasing lock "singleton_lock" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 474.702131] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Full set of CONF: {{(pid=62235) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 474.702277] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ******************************************************************************** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 474.702406] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Configuration options gathered from: {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 474.702546] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 474.702771] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 474.702905] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ================================================================================ {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 474.703130] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] allow_resize_to_same_host = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.703308] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] arq_binding_timeout = 300 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.703442] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] backdoor_port = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.703569] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] backdoor_socket = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.703735] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] block_device_allocate_retries = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.703897] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] block_device_allocate_retries_interval = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.704076] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cert = self.pem {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.704242] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.704410] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute_monitors = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.704580] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] config_dir = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.704750] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] config_drive_format = iso9660 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.704885] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.705057] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] config_source = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.705226] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] console_host = devstack {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.705390] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] control_exchange = nova {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.705554] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cpu_allocation_ratio = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.705714] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] daemon = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.705881] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] debug = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.706055] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] default_access_ip_network_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.706227] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] default_availability_zone = nova {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.706386] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] default_ephemeral_format = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.706550] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] default_green_pool_size = 1000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.706784] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.706949] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] default_schedule_zone = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.707157] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] disk_allocation_ratio = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.707329] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] enable_new_services = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.707508] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] enabled_apis = ['osapi_compute'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.707675] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] enabled_ssl_apis = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.707838] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] flat_injected = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.707998] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] force_config_drive = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.708172] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] force_raw_images = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.708339] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] graceful_shutdown_timeout = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.708503] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] heal_instance_info_cache_interval = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.708766] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] host = cpu-1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.708973] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.709178] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.709352] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.709568] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.709736] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instance_build_timeout = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.709898] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instance_delete_interval = 300 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.710108] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instance_format = [instance: %(uuid)s] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.710692] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instance_name_template = instance-%08x {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.710692] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instance_usage_audit = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.710692] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instance_usage_audit_period = month {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.710864] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.710969] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.711153] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] internal_service_availability_zone = internal {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.711313] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] key = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.711475] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] live_migration_retry_count = 30 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.711644] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_color = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.711810] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_config_append = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.711976] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.712153] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_dir = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.712313] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.712441] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_options = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.712616] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_rotate_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.712800] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_rotate_interval_type = days {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.712967] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] log_rotation_type = none {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.713110] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.713238] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.713406] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.713571] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.713699] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.713865] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] long_rpc_timeout = 1800 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.714035] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] max_concurrent_builds = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.714200] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] max_concurrent_live_migrations = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.714360] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] max_concurrent_snapshots = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.714521] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] max_local_block_devices = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.714678] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] max_logfile_count = 30 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.714836] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] max_logfile_size_mb = 200 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.714994] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] maximum_instance_delete_attempts = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.715177] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] metadata_listen = 0.0.0.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.715349] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] metadata_listen_port = 8775 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.715519] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] metadata_workers = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.715681] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] migrate_max_retries = -1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.715846] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] mkisofs_cmd = genisoimage {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.716061] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.716199] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] my_ip = 10.180.1.21 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.716363] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] network_allocate_retries = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.716540] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.716706] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.716867] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] osapi_compute_listen_port = 8774 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.717039] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] osapi_compute_unique_server_name_scope = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.717208] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] osapi_compute_workers = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.717370] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] password_length = 12 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.717530] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] periodic_enable = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.717688] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] periodic_fuzzy_delay = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.717854] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] pointer_model = usbtablet {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.718022] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] preallocate_images = none {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.718185] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] publish_errors = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.718312] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] pybasedir = /opt/stack/nova {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.718466] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ram_allocation_ratio = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.718622] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] rate_limit_burst = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.718785] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] rate_limit_except_level = CRITICAL {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.718942] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] rate_limit_interval = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.719114] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] reboot_timeout = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.719277] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] reclaim_instance_interval = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.719436] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] record = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.719602] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] reimage_timeout_per_gb = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.719764] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] report_interval = 120 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.719923] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] rescue_timeout = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.720094] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] reserved_host_cpus = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.720262] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] reserved_host_disk_mb = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.720440] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] reserved_host_memory_mb = 512 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.720609] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] reserved_huge_pages = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.720770] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] resize_confirm_window = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.720929] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] resize_fs_using_block_device = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.721149] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] resume_guests_state_on_host_boot = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.721283] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.721458] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] rpc_response_timeout = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.721618] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] run_external_periodic_tasks = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.721784] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] running_deleted_instance_action = reap {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.721944] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.722115] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] running_deleted_instance_timeout = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.722279] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler_instance_sync_interval = 120 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.722448] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_down_time = 720 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.722639] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] servicegroup_driver = db {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.722897] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] shell_completion = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.723157] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] shelved_offload_time = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.723339] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] shelved_poll_interval = 3600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.723513] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] shutdown_timeout = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.723676] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] source_is_ipv6 = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.723834] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ssl_only = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.724099] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.724274] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] sync_power_state_interval = 600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.724443] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] sync_power_state_pool_size = 1000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.724612] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] syslog_log_facility = LOG_USER {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.724771] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] tempdir = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.724930] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] timeout_nbd = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.725109] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] transport_url = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.725274] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] update_resources_interval = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.725436] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] use_cow_images = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.725594] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] use_eventlog = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.725753] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] use_journal = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.725909] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] use_json = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.726076] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] use_rootwrap_daemon = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.726236] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] use_stderr = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.726394] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] use_syslog = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.726553] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vcpu_pin_set = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.726721] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plugging_is_fatal = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.726889] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plugging_timeout = 300 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.727066] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] virt_mkfs = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.727232] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] volume_usage_poll_interval = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.727392] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] watch_log_file = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.727561] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] web = /usr/share/spice-html5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 474.727744] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.727912] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.728090] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.728265] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_concurrency.disable_process_locking = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.728562] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.728748] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.728916] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.729101] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.729277] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.729447] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.729630] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.auth_strategy = keystone {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.730161] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.compute_link_prefix = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.730161] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.730161] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.dhcp_domain = novalocal {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.730413] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.enable_instance_password = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.730523] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.glance_link_prefix = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.730666] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.730838] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.731011] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.instance_list_per_project_cells = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.731177] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.list_records_by_skipping_down_cells = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.731345] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.local_metadata_per_cell = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.731514] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.max_limit = 1000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.731686] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.metadata_cache_expiration = 15 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.731861] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.neutron_default_tenant_id = default {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.732043] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.response_validation = warn {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.732221] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.use_neutron_default_nets = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.732392] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.732557] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.732771] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.732959] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.733158] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.vendordata_dynamic_targets = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.733364] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.vendordata_jsonfile_path = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.733605] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.733809] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.backend = dogpile.cache.memcached {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.733981] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.backend_argument = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.734168] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.config_prefix = cache.oslo {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.734341] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.dead_timeout = 60.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.734509] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.debug_cache_backend = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.734674] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.enable_retry_client = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.734836] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.enable_socket_keepalive = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.735021] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.enabled = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.735184] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.enforce_fips_mode = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.735351] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.expiration_time = 600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.735518] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.hashclient_retry_attempts = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.735714] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.735914] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_dead_retry = 300 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.736094] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_password = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.736264] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.736430] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.736593] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_pool_maxsize = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.736756] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.736921] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_sasl_enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.737113] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.737284] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.737449] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.memcache_username = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.737617] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.proxies = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.737782] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.redis_db = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.737943] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.redis_password = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.738128] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.738308] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.738478] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.redis_server = localhost:6379 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.738641] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.redis_socket_timeout = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.738800] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.redis_username = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.738962] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.retry_attempts = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.739140] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.retry_delay = 0.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.739305] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.socket_keepalive_count = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.739468] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.socket_keepalive_idle = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.739631] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.socket_keepalive_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.739791] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.tls_allowed_ciphers = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.739950] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.tls_cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.740124] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.tls_certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.740317] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.tls_enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.740485] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cache.tls_keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.740658] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.740832] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.auth_type = password {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.740993] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.741182] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.741367] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.741539] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.741699] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.cross_az_attach = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.741860] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.debug = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.742028] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.endpoint_template = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.742198] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.http_retries = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.742361] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.742520] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.742714] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.os_region_name = RegionOne {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.742887] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.743061] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cinder.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.743240] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.743401] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.cpu_dedicated_set = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.743561] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.cpu_shared_set = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.743726] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.image_type_exclude_list = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.743888] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.744060] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.744226] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.744387] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.744556] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.744718] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.resource_provider_association_refresh = 300 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.744878] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.745048] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.shutdown_retry_interval = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.745231] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.745410] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] conductor.workers = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.745589] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] console.allowed_origins = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.745751] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] console.ssl_ciphers = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.745918] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] console.ssl_minimum_version = default {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.746099] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] consoleauth.enforce_session_timeout = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.746273] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] consoleauth.token_ttl = 600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.746443] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.746604] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.746768] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.746938] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.connect_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.747102] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.connect_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.747264] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.endpoint_override = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.747431] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.747592] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.747751] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.max_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.747907] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.min_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.748075] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.region_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.748238] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.retriable_status_codes = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.748395] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.service_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.748562] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.service_type = accelerator {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.748723] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.748878] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.status_code_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.749044] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.status_code_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.749207] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.749388] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.749552] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] cyborg.version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.749730] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.backend = sqlalchemy {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.749901] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.connection = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.750076] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.connection_debug = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.750254] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.connection_parameters = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.750444] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.connection_recycle_time = 3600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.750615] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.connection_trace = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.750779] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.db_inc_retry_interval = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.750943] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.db_max_retries = 20 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.751122] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.db_max_retry_interval = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.751289] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.db_retry_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.751454] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.max_overflow = 50 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.751615] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.max_pool_size = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.751777] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.max_retries = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.751947] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.752120] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.mysql_wsrep_sync_wait = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.752283] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.pool_timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.752446] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.retry_interval = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.752625] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.slave_connection = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.752807] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.sqlite_synchronous = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.752973] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] database.use_db_reconnect = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.753167] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.backend = sqlalchemy {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.753340] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.connection = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.753510] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.connection_debug = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.753677] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.connection_parameters = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.753840] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.connection_recycle_time = 3600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.754011] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.connection_trace = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.754182] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.db_inc_retry_interval = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.754351] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.db_max_retries = 20 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.754511] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.db_max_retry_interval = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.754674] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.db_retry_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.754834] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.max_overflow = 50 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.754996] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.max_pool_size = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.755173] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.max_retries = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.755343] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.755504] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.755661] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.pool_timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.755824] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.retry_interval = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.755981] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.slave_connection = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.756157] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] api_database.sqlite_synchronous = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.756337] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] devices.enabled_mdev_types = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.756514] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.756686] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.756851] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ephemeral_storage_encryption.enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.757027] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.757203] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.api_servers = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.757370] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.757532] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.757696] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.757853] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.connect_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.758017] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.connect_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.758183] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.debug = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.758349] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.default_trusted_certificate_ids = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.758514] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.enable_certificate_validation = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.758674] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.enable_rbd_download = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.758833] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.endpoint_override = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.758997] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.759173] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.759333] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.max_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.759490] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.min_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.759653] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.num_retries = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.759819] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.rbd_ceph_conf = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.759978] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.rbd_connect_timeout = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.760157] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.rbd_pool = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.760341] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.rbd_user = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.760507] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.region_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.760668] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.retriable_status_codes = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.760837] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.service_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.760989] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.service_type = image {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.761162] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.761335] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.status_code_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.761539] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.status_code_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.761652] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.761817] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.761977] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.verify_glance_signatures = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.762147] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] glance.version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.762317] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] guestfs.debug = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.762480] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] mks.enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.762848] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.763055] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] image_cache.manager_interval = 2400 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.763234] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] image_cache.precache_concurrency = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.763407] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] image_cache.remove_unused_base_images = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.763582] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.763753] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.763929] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] image_cache.subdirectory_name = _base {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.764117] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.api_max_retries = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.764286] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.api_retry_interval = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.764447] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.764610] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.auth_type = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.764771] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.764930] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.765106] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.765280] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.conductor_group = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.765458] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.connect_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.765614] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.connect_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.765770] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.endpoint_override = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.765934] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.766103] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.766267] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.max_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.766427] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.min_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.766592] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.peer_list = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.766750] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.region_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.766973] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.retriable_status_codes = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.767174] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.serial_console_state_timeout = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.767342] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.service_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.767516] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.service_type = baremetal {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.767679] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.shard = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.767844] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.768013] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.status_code_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.768180] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.status_code_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.768340] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.768520] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.768682] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ironic.version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.768864] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.769048] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] key_manager.fixed_key = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.769236] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.769397] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.barbican_api_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.769558] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.barbican_endpoint = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.769728] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.barbican_endpoint_type = public {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.769886] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.barbican_region_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.770053] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.770217] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.770411] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.770578] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.770738] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.770943] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.number_of_retries = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.771070] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.retry_delay = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.771237] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.send_service_user_token = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.771423] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.771587] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.771745] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.verify_ssl = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.771903] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican.verify_ssl_path = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.772083] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.772248] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.auth_type = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.772404] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.772561] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.772759] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.772931] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.773107] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.773273] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.773434] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] barbican_service_user.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.773600] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.approle_role_id = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.773759] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.approle_secret_id = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.773929] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.kv_mountpoint = secret {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.774097] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.kv_path = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.774265] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.kv_version = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.774426] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.namespace = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.774584] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.root_token_id = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.774739] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.ssl_ca_crt_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.774903] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.timeout = 60.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.775076] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.use_ssl = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.775247] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.775416] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.775579] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.auth_type = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.775739] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.775893] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.776066] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.776223] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.connect_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.776379] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.connect_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.776534] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.endpoint_override = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.776691] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.776843] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.776996] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.max_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.777163] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.min_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.777319] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.region_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.777474] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.retriable_status_codes = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.777628] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.service_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.777791] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.service_type = identity {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.777951] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.778117] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.status_code_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.778277] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.status_code_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.778434] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.778612] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.778791] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] keystone.version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.779022] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.connection_uri = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.779185] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.cpu_mode = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.779354] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.779526] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.cpu_models = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.779697] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.cpu_power_governor_high = performance {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.779865] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.780037] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.cpu_power_management = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.780213] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.780407] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.device_detach_attempts = 8 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.780579] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.device_detach_timeout = 20 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.780751] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.disk_cachemodes = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.780906] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.disk_prefix = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.781080] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.enabled_perf_events = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.781250] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.file_backed_memory = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.781446] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.gid_maps = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.781611] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.hw_disk_discard = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.781766] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.hw_machine_type = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.781936] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.images_rbd_ceph_conf = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.782118] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.782284] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.782453] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.images_rbd_glance_store_name = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.782640] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.images_rbd_pool = rbd {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.782822] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.images_type = default {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.782980] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.images_volume_group = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.783157] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.inject_key = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.783321] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.inject_partition = -2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.783486] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.inject_password = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.783643] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.iscsi_iface = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.783804] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.iser_use_multipath = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.783969] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.784148] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.784312] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_downtime = 500 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.784477] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.784637] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.784798] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_inbound_addr = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.784959] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.785139] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.785303] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_scheme = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.785487] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_timeout_action = abort {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.785653] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_tunnelled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.785813] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_uri = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.785973] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.live_migration_with_native_tls = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.786142] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.max_queues = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.786305] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.786539] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.786697] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.nfs_mount_options = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.786992] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.787180] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.787357] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.787520] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.787684] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.787846] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.num_pcie_ports = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.788067] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.788259] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.pmem_namespaces = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.788449] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.quobyte_client_cfg = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.788743] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.788920] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.789100] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.789269] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.789432] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rbd_secret_uuid = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.789590] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rbd_user = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.789754] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.789925] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.790095] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rescue_image_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.790268] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rescue_kernel_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.790446] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rescue_ramdisk_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.790618] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.790777] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.rx_queue_size = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.790945] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.smbfs_mount_options = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.791238] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.791413] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.snapshot_compression = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.791576] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.snapshot_image_format = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.791789] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.791957] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.sparse_logical_volumes = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.792133] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.swtpm_enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.792303] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.swtpm_group = tss {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.792471] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.swtpm_user = tss {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.792656] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.sysinfo_serial = unique {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.792828] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.tb_cache_size = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.792986] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.tx_queue_size = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.793167] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.uid_maps = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.793331] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.use_virtio_for_bridges = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.793501] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.virt_type = kvm {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.793669] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.volume_clear = zero {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.793833] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.volume_clear_size = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.793998] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.volume_use_multipath = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.794171] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.vzstorage_cache_path = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.794340] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.794509] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.794675] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.794842] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.795127] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.795305] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.vzstorage_mount_user = stack {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.795477] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.795651] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.795823] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.auth_type = password {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.795985] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.796158] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.796321] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.796481] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.connect_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.796639] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.connect_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.796807] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.default_floating_pool = public {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.796965] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.endpoint_override = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.797139] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.extension_sync_interval = 600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.797303] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.http_retries = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.797463] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.797619] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.797774] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.max_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.797941] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.798108] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.min_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.798277] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.ovs_bridge = br-int {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.798442] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.physnets = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.798607] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.region_name = RegionOne {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.798764] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.retriable_status_codes = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.798927] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.service_metadata_proxy = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.799096] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.service_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.799265] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.service_type = network {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.799428] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.799585] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.status_code_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.799741] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.status_code_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.799898] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.800086] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.800252] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] neutron.version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.800445] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] notifications.bdms_in_notifications = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.800624] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] notifications.default_level = INFO {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.800796] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] notifications.notification_format = unversioned {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.800960] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] notifications.notify_on_state_change = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.801148] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.801325] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] pci.alias = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.801496] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] pci.device_spec = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.801659] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] pci.report_in_placement = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.801828] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.801998] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.auth_type = password {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.802175] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.802335] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.802492] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.802672] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.802841] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.connect_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.803008] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.connect_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.803172] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.default_domain_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.803329] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.default_domain_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.803487] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.domain_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.803642] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.domain_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.803800] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.endpoint_override = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.803960] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.804128] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.804288] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.max_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.804440] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.min_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.804603] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.password = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.804759] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.project_domain_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.804922] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.project_domain_name = Default {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.805104] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.project_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.805277] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.project_name = service {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.805446] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.region_name = RegionOne {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.805608] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.retriable_status_codes = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.805767] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.service_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.805934] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.service_type = placement {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.806107] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.806269] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.status_code_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.806432] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.status_code_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.806590] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.system_scope = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.806746] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.806901] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.trust_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.807068] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.user_domain_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.807236] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.user_domain_name = Default {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.807396] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.user_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.807569] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.username = nova {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.807750] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.807910] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] placement.version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.808103] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.cores = 20 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.808274] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.count_usage_from_placement = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.808447] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.808622] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.injected_file_content_bytes = 10240 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.808789] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.injected_file_path_length = 255 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.808955] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.injected_files = 5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.809135] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.instances = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.809306] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.key_pairs = 100 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.809474] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.metadata_items = 128 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.809641] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.ram = 51200 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.809804] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.recheck_quota = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.809971] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.server_group_members = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.810150] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] quota.server_groups = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.810349] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.810529] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.810696] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.image_metadata_prefilter = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.810859] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.811032] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.max_attempts = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.811201] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.max_placement_results = 1000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.811375] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.811550] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.811715] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.811887] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] scheduler.workers = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.812074] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.812245] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.812422] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.812602] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.812787] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.812952] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.813130] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.813321] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.813496] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.host_subset_size = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.813661] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.813823] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.813986] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.814169] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.isolated_hosts = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.814340] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.isolated_images = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.814505] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.814665] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.814830] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.814991] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.pci_in_placement = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.815171] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.815334] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.815500] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.815660] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.815824] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.815988] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.816162] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.track_instance_changes = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.816338] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.816509] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] metrics.required = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.816675] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] metrics.weight_multiplier = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.816840] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.817013] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] metrics.weight_setting = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.817337] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.817515] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] serial_console.enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.817691] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] serial_console.port_range = 10000:20000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.817864] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.818047] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.818222] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] serial_console.serialproxy_port = 6083 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.818391] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.818565] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.auth_type = password {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.818725] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.818881] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.819051] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.819216] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.819374] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.819545] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.send_service_user_token = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.819708] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.819880] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] service_user.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.820062] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.agent_enabled = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.820231] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.820564] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.820767] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.820940] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.html5proxy_port = 6082 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.821118] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.image_compression = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.821283] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.jpeg_compression = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.821473] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.playback_compression = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.821602] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.require_secure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.821768] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.server_listen = 127.0.0.1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.821935] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.822109] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.streaming_mode = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.822266] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] spice.zlib_compression = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.822431] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] upgrade_levels.baseapi = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.822616] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] upgrade_levels.compute = auto {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.822800] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] upgrade_levels.conductor = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.822963] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] upgrade_levels.scheduler = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.823145] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.823338] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.823473] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.823630] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.823789] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.823947] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.824116] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.824281] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.824441] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vendordata_dynamic_auth.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.824617] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.api_retry_count = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.824776] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.ca_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.824945] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.825124] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.cluster_name = testcl1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.825291] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.connection_pool_size = 10 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.825451] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.console_delay_seconds = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.825616] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.datastore_regex = ^datastore.* {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.825817] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.825989] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.host_password = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.826168] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.host_port = 443 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.826339] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.host_username = administrator@vsphere.local {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.826506] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.insecure = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.826666] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.integration_bridge = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.826829] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.maximum_objects = 100 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.826986] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.pbm_default_policy = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.827159] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.pbm_enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.827320] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.pbm_wsdl_location = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.827486] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.827644] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.serial_port_proxy_uri = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.827799] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.serial_port_service_uri = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.827960] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.task_poll_interval = 0.5 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.828143] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.use_linked_clone = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.828312] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.vnc_keymap = en-us {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.828476] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.vnc_port = 5900 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.828638] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vmware.vnc_port_total = 10000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.828821] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.auth_schemes = ['none'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.828994] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.829298] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.829486] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.829655] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.novncproxy_port = 6080 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.829829] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.server_listen = 127.0.0.1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.829998] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.830217] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.vencrypt_ca_certs = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.830426] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.vencrypt_client_cert = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.830595] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vnc.vencrypt_client_key = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.830770] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.830935] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.disable_deep_image_inspection = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.831111] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.831279] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.831439] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.831598] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.disable_rootwrap = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.831756] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.enable_numa_live_migration = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.831915] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.832086] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.832282] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.832402] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.libvirt_disable_apic = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.832559] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.832765] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.832939] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.833119] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.833285] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.833447] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.833608] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.833767] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.833925] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.834101] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.834288] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.834460] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.client_socket_timeout = 900 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.834627] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.default_pool_size = 1000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.834793] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.keep_alive = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.834960] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.max_header_line = 16384 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.835135] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.835297] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.ssl_ca_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.835458] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.ssl_cert_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.835617] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.ssl_key_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.835859] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.tcp_keepidle = 600 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.836075] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.836250] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] zvm.ca_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.836412] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] zvm.cloud_connector_url = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.836693] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.836869] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] zvm.reachable_timeout = 300 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.837060] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.enforce_new_defaults = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.837438] env[62235]: WARNING oslo_config.cfg [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 474.837624] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.enforce_scope = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.837803] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.policy_default_rule = default {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.837988] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.838182] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.policy_file = policy.yaml {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.838350] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.838513] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.838671] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.838828] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.838986] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.839168] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.839341] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.839515] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.connection_string = messaging:// {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.839680] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.enabled = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.839847] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.es_doc_type = notification {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.840018] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.es_scroll_size = 10000 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.840191] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.es_scroll_time = 2m {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.840377] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.filter_error_trace = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.840557] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.hmac_keys = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.840726] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.sentinel_service_name = mymaster {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.840894] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.socket_timeout = 0.1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.841066] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.trace_requests = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.841230] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler.trace_sqlalchemy = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.841492] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler_jaeger.process_tags = {} {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.841682] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler_jaeger.service_name_prefix = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.841805] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] profiler_otlp.service_name_prefix = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.841975] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] remote_debug.host = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.842150] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] remote_debug.port = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.842377] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.842585] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.842789] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.842962] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.843142] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.843305] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.843468] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.843630] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.843795] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.843965] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.844138] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.844311] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.844481] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.844652] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.844821] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.844984] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.845162] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.845337] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.845501] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.845664] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.845829] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.845993] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.846169] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.846361] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.846540] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.846703] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.846868] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.847041] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.847213] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.847381] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.ssl = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.847556] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.847725] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.847888] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.848068] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.848243] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.848405] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.848593] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.848763] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_notifications.retry = -1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.848945] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.849135] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.849310] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.auth_section = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.849476] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.auth_type = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.849634] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.cafile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.849789] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.certfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.849951] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.collect_timing = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.850124] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.connect_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.850328] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.connect_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.850483] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.endpoint_id = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.850645] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.endpoint_override = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.850807] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.insecure = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.850965] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.keyfile = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.851133] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.max_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.851291] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.min_version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.851445] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.region_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.851602] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.retriable_status_codes = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.851786] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.service_name = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.851913] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.service_type = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.852086] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.split_loggers = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.852249] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.status_code_retries = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.852408] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.status_code_retry_delay = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.852566] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.timeout = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.852763] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.valid_interfaces = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.852931] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_limit.version = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.853114] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_reports.file_event_handler = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.853287] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.853449] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] oslo_reports.log_dir = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.853617] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.853775] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.853930] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.854106] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.854273] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.854431] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.854599] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.854757] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_ovs_privileged.group = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.854914] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.855085] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.855249] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.855406] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] vif_plug_ovs_privileged.user = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.855577] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.855754] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.855925] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.856106] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.856280] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.856459] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.856627] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.856790] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.856967] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.857153] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_ovs.isolate_vif = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.857323] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.857492] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.857658] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.857825] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.857987] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] os_vif_ovs.per_port_bridge = False {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.858175] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] privsep_osbrick.capabilities = [21] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.858355] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] privsep_osbrick.group = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.858526] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] privsep_osbrick.helper_command = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.858690] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.858855] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.859020] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] privsep_osbrick.user = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.859202] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.859361] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] nova_sys_admin.group = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.859520] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] nova_sys_admin.helper_command = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.859683] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.859844] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.859999] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] nova_sys_admin.user = None {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 474.860139] env[62235]: DEBUG oslo_service.service [None req-afe4d80a-9044-44a1-bece-f236443ae840 None None] ******************************************************************************** {{(pid=62235) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 474.860640] env[62235]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 475.364293] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Getting list of instances from cluster (obj){ [ 475.364293] env[62235]: value = "domain-c8" [ 475.364293] env[62235]: _type = "ClusterComputeResource" [ 475.364293] env[62235]: } {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 475.365436] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92788f4c-1250-4d78-8f91-d9aba07acb25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.374238] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Got total of 0 instances {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 475.374764] env[62235]: WARNING nova.virt.vmwareapi.driver [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 475.375241] env[62235]: INFO nova.virt.node [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Generated node identity 108a4390-ff68-4048-b61d-c7a9614ddc4f [ 475.375483] env[62235]: INFO nova.virt.node [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Wrote node identity 108a4390-ff68-4048-b61d-c7a9614ddc4f to /opt/stack/data/n-cpu-1/compute_id [ 475.879291] env[62235]: WARNING nova.compute.manager [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Compute nodes ['108a4390-ff68-4048-b61d-c7a9614ddc4f'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 476.885104] env[62235]: INFO nova.compute.manager [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 477.890881] env[62235]: WARNING nova.compute.manager [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 477.891315] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 477.891379] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 477.891581] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.891742] env[62235]: DEBUG nova.compute.resource_tracker [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 477.892669] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f229453e-30c0-4818-9520-27b30e6cb96a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.901932] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a781ff-e60d-4cf8-a347-de141667ec89 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.916653] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41461768-ff55-4078-8e0e-f43af7f22678 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.922575] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466d7720-ed69-46b4-a4ae-965b01c9991f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.950334] env[62235]: DEBUG nova.compute.resource_tracker [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181499MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 477.950551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 477.950782] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 478.453875] env[62235]: WARNING nova.compute.resource_tracker [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] No compute node record for cpu-1:108a4390-ff68-4048-b61d-c7a9614ddc4f: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 108a4390-ff68-4048-b61d-c7a9614ddc4f could not be found. [ 478.957406] env[62235]: INFO nova.compute.resource_tracker [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 108a4390-ff68-4048-b61d-c7a9614ddc4f [ 480.466266] env[62235]: DEBUG nova.compute.resource_tracker [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 480.466597] env[62235]: DEBUG nova.compute.resource_tracker [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 480.615147] env[62235]: INFO nova.scheduler.client.report [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] [req-f68a3d1c-91fe-4f13-8bd7-8311e19e62c9] Created resource provider record via placement API for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 480.631831] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b5da8b-6172-43a9-8114-bef23848ff06 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.639359] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9eec898-179c-4538-871d-0c91c2926d09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.669438] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfd0d44-622f-41a7-b841-5fc118701ff3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.676201] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1a294a-8e6c-446d-bded-ddda95f48386 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.688613] env[62235]: DEBUG nova.compute.provider_tree [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.228979] env[62235]: DEBUG nova.scheduler.client.report [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Updated inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 481.228979] env[62235]: DEBUG nova.compute.provider_tree [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 0 to 1 during operation: update_inventory {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 481.228979] env[62235]: DEBUG nova.compute.provider_tree [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.280088] env[62235]: DEBUG nova.compute.provider_tree [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 1 to 2 during operation: update_traits {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 481.785768] env[62235]: DEBUG nova.compute.resource_tracker [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 481.785768] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.833s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 481.785768] env[62235]: DEBUG nova.service [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Creating RPC server for service compute {{(pid=62235) start /opt/stack/nova/nova/service.py:186}} [ 481.799241] env[62235]: DEBUG nova.service [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] Join ServiceGroup membership for this service compute {{(pid=62235) start /opt/stack/nova/nova/service.py:203}} [ 481.799619] env[62235]: DEBUG nova.servicegroup.drivers.db [None req-3d7154a7-e51e-4133-ab21-cf19317949ec None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62235) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 510.100306] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "eb6cc4b1-9236-417c-bdda-35afff032da5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.100306] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "eb6cc4b1-9236-417c-bdda-35afff032da5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.109209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.109417] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.602806] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 510.612209] env[62235]: DEBUG nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.150047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.150417] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.151977] env[62235]: INFO nova.compute.claims [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 511.159217] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.236098] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b563b2-031a-4dad-9840-aa9779f3f888 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.246502] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d419513-63bd-4795-b596-1f15e590e5b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.281327] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97c57f8-656a-4c12-99d1-e82633fc47ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.290186] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341faa4a-9a26-49c5-aeb4-a9b31ca4be08 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.303363] env[62235]: DEBUG nova.compute.provider_tree [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 512.802422] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 512.808410] env[62235]: DEBUG nova.scheduler.client.report [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 513.307121] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Getting list of instances from cluster (obj){ [ 513.307121] env[62235]: value = "domain-c8" [ 513.307121] env[62235]: _type = "ClusterComputeResource" [ 513.307121] env[62235]: } {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 513.308311] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1a3a7c-79c7-49c1-b0ce-944a2197da33 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.316398] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.317000] env[62235]: DEBUG nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 513.326114] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.167s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.327883] env[62235]: INFO nova.compute.claims [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.330463] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Got total of 0 instances {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 513.330612] env[62235]: WARNING nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] While synchronizing instance power states, found 1 instances in the database and 0 instances on the hypervisor. [ 513.330735] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Triggering sync for uuid d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0 {{(pid=62235) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 513.333412] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.333697] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 513.334022] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Getting list of instances from cluster (obj){ [ 513.334022] env[62235]: value = "domain-c8" [ 513.334022] env[62235]: _type = "ClusterComputeResource" [ 513.334022] env[62235]: } {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 513.336248] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9816d8c4-8623-4921-b214-62629b7a6d8e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.351765] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Got total of 0 instances {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 513.837151] env[62235]: DEBUG nova.compute.utils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.839624] env[62235]: DEBUG nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Not allocating networking since 'none' was specified. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 514.344171] env[62235]: DEBUG nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 514.438798] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.439074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.459776] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a045469e-a94f-41f7-ba49-66bd2e9808fa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.469978] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9662a1b6-2d8d-4b14-8f83-3afd691df3cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.511381] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb6aca6-cadb-4634-8a9e-04e4e39ebcca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.519604] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9985066f-40e5-45ee-86d6-7a9d1af0c08f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.533780] env[62235]: DEBUG nova.compute.provider_tree [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 514.944160] env[62235]: DEBUG nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.041270] env[62235]: DEBUG nova.scheduler.client.report [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 515.218845] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "9617c521-d878-4749-a9b6-220ec51a2fc1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.219186] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "9617c521-d878-4749-a9b6-220ec51a2fc1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.268677] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "df5f7759-c872-4044-9901-200eb7933543" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.268914] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "df5f7759-c872-4044-9901-200eb7933543" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.304998] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "dd37deb6-c5d8-44d5-8c2a-319ffce9f344" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.304998] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "dd37deb6-c5d8-44d5-8c2a-319ffce9f344" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.360340] env[62235]: DEBUG nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 515.484052] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.550238] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.550758] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 515.555147] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.071s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.556556] env[62235]: INFO nova.compute.claims [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.562018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Acquiring lock "6d33eefc-b822-40fa-8b68-adb944e7dc7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.562234] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Lock "6d33eefc-b822-40fa-8b68-adb944e7dc7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.594764] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.595112] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.595262] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.595449] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.595654] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.595766] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.595990] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.596208] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.596802] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.597368] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.597500] env[62235]: DEBUG nova.virt.hardware [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.599276] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1301c4-af07-4b0e-8933-47bbadd7e193 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.611223] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050ee456-d0aa-4a5e-80a7-a1c8cc6a8d02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.632700] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e38215-63b3-4618-b68c-84d663252118 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.661146] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 515.672568] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.675349] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7c28eba-7925-439a-8bcb-f4a221896b3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.687822] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Created folder: OpenStack in parent group-v4. [ 515.688480] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating folder: Project (e1eec05f97f34c40ab85d2e16420968e). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.688480] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-213e35eb-001b-4dd8-bc8b-80a6981dc56d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.701683] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.701990] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.710958] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Created folder: Project (e1eec05f97f34c40ab85d2e16420968e) in parent group-v273362. [ 515.710958] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating folder: Instances. Parent ref: group-v273363. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.710958] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75a5c5de-7f2f-4175-82bd-2cb5ffb8fb29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.721579] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Created folder: Instances in parent group-v273363. [ 515.721845] env[62235]: DEBUG oslo.service.loopingcall [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 515.722142] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 515.722391] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4900b31f-2ac7-4197-9c32-b146a834fffe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.740903] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.752774] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 515.752774] env[62235]: value = "task-1271403" [ 515.752774] env[62235]: _type = "Task" [ 515.752774] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.765019] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271403, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.771618] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.808951] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.071028] env[62235]: DEBUG nova.compute.utils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 516.071028] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.075130] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 516.077114] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 516.206558] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.266725] env[62235]: DEBUG nova.policy [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '242e31f81b7944c3900af425422d1157', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd90b53cbac954198863c36ef627925b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 516.289485] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271403, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.298018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.299483] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.340957] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.582738] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 516.613919] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.694149] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "d386f90d-25c3-4843-a91f-a092d27a6b5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.694355] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "d386f90d-25c3-4843-a91f-a092d27a6b5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.747502] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.749898] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524e425f-95b6-438d-a5b1-fa8b21725f6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.758942] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f75d12-69af-451d-baca-a1fd15bef41c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.802539] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e4cb51-0e8d-4914-9460-472a8574e78b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.817184] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271403, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.823884] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520c80e7-9dc7-4647-bff0-a394bdbdd73e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.838185] env[62235]: DEBUG nova.compute.provider_tree [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.199439] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.313165] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271403, 'name': CreateVM_Task, 'duration_secs': 1.315563} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.313165] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 517.315167] env[62235]: DEBUG oslo_vmware.service [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c40ed9-6060-441c-a9bf-7068e14b6b10 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.320346] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.321043] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.321231] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 517.321411] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74261bc8-1f86-4e4d-8f6c-3059f9a9eab9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.325958] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 517.325958] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520bfccd-ef65-1298-d15d-4eef705fe192" [ 517.325958] env[62235]: _type = "Task" [ 517.325958] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.337200] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520bfccd-ef65-1298-d15d-4eef705fe192, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.349650] env[62235]: DEBUG nova.scheduler.client.report [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.598656] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 517.659626] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 517.659866] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 517.660031] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 517.660215] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 517.660356] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 517.660536] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 517.660947] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 517.660947] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 517.661116] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 517.661272] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 517.661468] env[62235]: DEBUG nova.virt.hardware [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 517.662419] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8738cc02-d042-4381-9d3a-7907b4812389 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.673666] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4dd1a8-dd39-462a-9445-0699ff4dbbcd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.740337] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.742818] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Successfully created port: d9ab7d0d-d72b-40df-b207-fa6d04040e3f {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 517.842533] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.842918] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 517.846746] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.846746] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.846746] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 517.846746] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68c5b1d6-5338-4956-95ff-dd642216d597 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.861577] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.862188] env[62235]: DEBUG nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.872026] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.570s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.872026] env[62235]: INFO nova.compute.claims [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 517.872026] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 517.872026] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 517.877545] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b9fcea-95cf-422e-9e46-3b3d546dabeb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.895407] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1c0dd51-0348-4e94-a89e-d39e49d84db1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.901734] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 517.901734] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bea71d-a743-9c62-a235-394ce5bb0338" [ 517.901734] env[62235]: _type = "Task" [ 517.901734] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.911525] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bea71d-a743-9c62-a235-394ce5bb0338, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.382099] env[62235]: DEBUG nova.compute.utils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.383979] env[62235]: DEBUG nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Not allocating networking since 'none' was specified. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 518.416384] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Preparing fetch location {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 518.416655] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating directory with path [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 518.416914] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80997f83-a19b-413a-800b-5c7d667b2144 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.439927] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Created directory with path [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 518.439927] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Fetch image to [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 518.439927] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Downloading image file data 02c64327-1eca-4d55-9f01-62fa8b2a2334 to [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk on the data store datastore2 {{(pid=62235) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 518.440607] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b22ae12-85a7-455c-b98d-4a15ef9c3815 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.449651] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cbe454-52cd-4887-a815-f17bf60c105a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.461822] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10190389-eb05-4dae-8a5d-b9f62b7f0d2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.510323] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf28985-12ab-4e96-9b6f-b97717a7c6a6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.517137] env[62235]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-02df394b-ffe8-47ca-baac-8f310210d893 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.604129] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Downloading image file data 02c64327-1eca-4d55-9f01-62fa8b2a2334 to the data store datastore2 {{(pid=62235) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 518.695464] env[62235]: DEBUG oslo_vmware.rw_handles [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62235) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 518.889378] env[62235]: DEBUG nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 519.121254] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7426f08e-01a3-49fc-b2a7-60f36d6c53a3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.134460] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8fda0c-b72c-49b2-bdb1-3ebc382c6eb7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.178242] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5f0995-3793-4ba8-9afa-736e97a5b4f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.186374] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0675efa-2338-4180-b528-071db89aefa3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.203808] env[62235]: DEBUG nova.compute.provider_tree [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.345392] env[62235]: DEBUG oslo_vmware.rw_handles [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Completed reading data from the image iterator. {{(pid=62235) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 519.345621] env[62235]: DEBUG oslo_vmware.rw_handles [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 519.529701] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Downloaded image file data 02c64327-1eca-4d55-9f01-62fa8b2a2334 to vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk on the data store datastore2 {{(pid=62235) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 519.529701] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Caching image {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 519.529701] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copying Virtual Disk [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk to [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 519.531043] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f63b57b-524c-4329-be7d-1360f8468043 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.540274] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 519.540274] env[62235]: value = "task-1271404" [ 519.540274] env[62235]: _type = "Task" [ 519.540274] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.551867] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.709510] env[62235]: DEBUG nova.scheduler.client.report [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.898027] env[62235]: DEBUG nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.941314] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.941682] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.941869] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.944058] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.944478] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.944478] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.944647] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.944855] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.945045] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.945207] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.945381] env[62235]: DEBUG nova.virt.hardware [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.946308] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b918b425-2dc6-4ed8-84a1-05fa93108c9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.960508] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7d4873-b265-4870-ab37-6a65f42d4b66 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.977314] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 519.984038] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Creating folder: Project (3d32fd385c724bd8848697690148883c). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 519.984038] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-223652c4-7e28-4336-9dbc-fde4918283ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.994508] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Created folder: Project (3d32fd385c724bd8848697690148883c) in parent group-v273362. [ 519.994656] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Creating folder: Instances. Parent ref: group-v273366. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 519.994864] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07363247-1dd2-419d-b7cd-5f783c6d3d4d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.006468] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Created folder: Instances in parent group-v273366. [ 520.006913] env[62235]: DEBUG oslo.service.loopingcall [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.007320] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 520.007558] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8aee868-627a-4f9a-8a92-317fd99fe7b3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.031680] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 520.031680] env[62235]: value = "task-1271407" [ 520.031680] env[62235]: _type = "Task" [ 520.031680] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.039352] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271407, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.049336] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271404, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.208034] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Acquiring lock "db07ab9b-13a6-42e5-bc60-1a1919696f0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.208034] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Lock "db07ab9b-13a6-42e5-bc60-1a1919696f0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.215844] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.215844] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.220186] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.919s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.223568] env[62235]: INFO nova.compute.claims [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.545168] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271407, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.554174] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271404, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730252} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.554174] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copied Virtual Disk [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk to [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 520.554174] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleting the datastore file [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 520.558197] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b3f02ed-9589-40f7-a0ea-c4b71daa84cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.565572] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 520.565572] env[62235]: value = "task-1271408" [ 520.565572] env[62235]: _type = "Task" [ 520.565572] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.580028] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271408, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.713855] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.725613] env[62235]: DEBUG nova.compute.utils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.731683] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 520.731683] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 520.847429] env[62235]: DEBUG nova.policy [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94bf8427bb65448ea2dc62d0dd05e519', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b19fc8b5c04d4311a1f5774aade4e0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.049130] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271407, 'name': CreateVM_Task, 'duration_secs': 0.525271} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.049319] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 521.049742] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.049888] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.050235] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 521.050494] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7afb6ad9-258d-4302-ab1a-39573f41d766 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.061764] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 521.061764] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b5886e-1437-0430-0034-effdbcfad205" [ 521.061764] env[62235]: _type = "Task" [ 521.061764] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.081908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.081908] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 521.081908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.081908] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271408, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.04086} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.081908] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 521.082154] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Moving file from [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0/02c64327-1eca-4d55-9f01-62fa8b2a2334 to [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334. {{(pid=62235) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 521.082154] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-9fe94ffb-13eb-41b2-854c-e1ca43db0cdf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.090593] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 521.090593] env[62235]: value = "task-1271409" [ 521.090593] env[62235]: _type = "Task" [ 521.090593] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.098749] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271409, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.230088] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.246906] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.316202] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Successfully created port: 0a437e0e-ed26-46ff-9867-052e8eca1eb2 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.477018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328573df-8b58-4dd3-bdc6-85e93eb3ee66 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.487159] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4aefb87-54c4-4aa0-9b7c-1266b3e54d48 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.527942] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271c78fe-b5c1-4c49-b7c5-2a97f9b74107 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.542124] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06183bf-d52f-4554-acad-137d11ace705 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.566063] env[62235]: DEBUG nova.compute.provider_tree [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.607461] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271409, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.055854} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.607891] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] File moved {{(pid=62235) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 521.608211] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Cleaning up location [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 521.608481] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleting the datastore file [datastore2] vmware_temp/d20fbba9-87cf-4a39-9cb0-8c6942a82df0 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 521.608837] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a1b9d24-2d26-44ac-8bec-7bfe501e21d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.617810] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 521.617810] env[62235]: value = "task-1271410" [ 521.617810] env[62235]: _type = "Task" [ 521.617810] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.627092] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.069899] env[62235]: DEBUG nova.scheduler.client.report [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.129914] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045512} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.129914] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 522.130480] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59a31edc-d4ba-4a85-a23b-08336cea4da8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.137745] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 522.137745] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2650a-bf0a-48ed-78d0-5efc9117c796" [ 522.137745] env[62235]: _type = "Task" [ 522.137745] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.146059] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2650a-bf0a-48ed-78d0-5efc9117c796, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.242999] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.268984] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.269368] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.269448] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.269621] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.269759] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.269931] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.270273] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.270341] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.270518] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.270651] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.270833] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.271981] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556fa3ea-6874-425a-9652-109c22089e44 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.280343] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5078fea6-ca65-43bb-878c-84a62c93ce7f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.481879] env[62235]: ERROR nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. [ 522.481879] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 522.481879] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.481879] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 522.481879] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.481879] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 522.481879] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.481879] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 522.481879] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.481879] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 522.481879] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.481879] env[62235]: ERROR nova.compute.manager raise self.value [ 522.481879] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.481879] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 522.481879] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.481879] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 522.482449] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.482449] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 522.482449] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. [ 522.482449] env[62235]: ERROR nova.compute.manager [ 522.482596] env[62235]: Traceback (most recent call last): [ 522.482652] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 522.482652] env[62235]: listener.cb(fileno) [ 522.482652] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.482652] env[62235]: result = function(*args, **kwargs) [ 522.482652] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.482652] env[62235]: return func(*args, **kwargs) [ 522.482819] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.482819] env[62235]: raise e [ 522.482819] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.482819] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 522.482819] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.482819] env[62235]: created_port_ids = self._update_ports_for_instance( [ 522.482819] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.482819] env[62235]: with excutils.save_and_reraise_exception(): [ 522.482819] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.482819] env[62235]: self.force_reraise() [ 522.482819] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.482819] env[62235]: raise self.value [ 522.482819] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.482819] env[62235]: updated_port = self._update_port( [ 522.482819] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.482819] env[62235]: _ensure_no_port_binding_failure(port) [ 522.482819] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.482819] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 522.482819] env[62235]: nova.exception.PortBindingFailed: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. [ 522.482819] env[62235]: Removing descriptor: 16 [ 522.485419] env[62235]: ERROR nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Traceback (most recent call last): [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] yield resources [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self.driver.spawn(context, instance, image_meta, [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] vm_ref = self.build_virtual_machine(instance, [ 522.485419] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] for vif in network_info: [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return self._sync_wrapper(fn, *args, **kwargs) [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self.wait() [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self[:] = self._gt.wait() [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return self._exit_event.wait() [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.485999] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] result = hub.switch() [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return self.greenlet.switch() [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] result = function(*args, **kwargs) [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return func(*args, **kwargs) [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] raise e [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] nwinfo = self.network_api.allocate_for_instance( [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] created_port_ids = self._update_ports_for_instance( [ 522.486344] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] with excutils.save_and_reraise_exception(): [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self.force_reraise() [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] raise self.value [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] updated_port = self._update_port( [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] _ensure_no_port_binding_failure(port) [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] raise exception.PortBindingFailed(port_id=port['id']) [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] nova.exception.PortBindingFailed: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. [ 522.486823] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] [ 522.487318] env[62235]: INFO nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Terminating instance [ 522.489304] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.489859] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquired lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.490034] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 522.581024] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.581024] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.584620] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.244s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.587013] env[62235]: INFO nova.compute.claims [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.650984] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2650a-bf0a-48ed-78d0-5efc9117c796, 'name': SearchDatastore_Task, 'duration_secs': 0.008692} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.651781] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.652302] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 522.652954] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.653388] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 522.653897] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efde8e29-6834-4228-9e51-4f518b27f9a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.657661] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23531ddb-2794-4667-ac9f-20ac93655a6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.668636] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 522.668636] env[62235]: value = "task-1271411" [ 522.668636] env[62235]: _type = "Task" [ 522.668636] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.669932] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 522.669979] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 522.675521] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecaf2f84-c23a-4f7b-9ba1-ffc7676aa0f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.683098] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.684842] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 522.684842] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520d0e4f-575b-8b83-6fac-f830b1a7706d" [ 522.684842] env[62235]: _type = "Task" [ 522.684842] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.694404] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520d0e4f-575b-8b83-6fac-f830b1a7706d, 'name': SearchDatastore_Task, 'duration_secs': 0.008047} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.695123] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-099784b0-e607-4086-8c5b-3906789f712d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.704621] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 522.704621] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b11235-69ec-7398-16ae-3572140981c5" [ 522.704621] env[62235]: _type = "Task" [ 522.704621] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.713433] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b11235-69ec-7398-16ae-3572140981c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.034369] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.090768] env[62235]: DEBUG nova.compute.utils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.096831] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.097747] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 523.160579] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. [ 523.160579] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.160579] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.160579] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.160579] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.160579] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.160579] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.160579] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.160579] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.160579] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 523.160579] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.160579] env[62235]: ERROR nova.compute.manager raise self.value [ 523.160579] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.160579] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.160579] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.160579] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.161165] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.161165] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.161165] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. [ 523.161165] env[62235]: ERROR nova.compute.manager [ 523.161165] env[62235]: Traceback (most recent call last): [ 523.161165] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.161165] env[62235]: listener.cb(fileno) [ 523.161165] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.161165] env[62235]: result = function(*args, **kwargs) [ 523.161165] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.161165] env[62235]: return func(*args, **kwargs) [ 523.161165] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.161165] env[62235]: raise e [ 523.161165] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.161165] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 523.161165] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.161165] env[62235]: created_port_ids = self._update_ports_for_instance( [ 523.161165] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.161165] env[62235]: with excutils.save_and_reraise_exception(): [ 523.161165] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.161165] env[62235]: self.force_reraise() [ 523.161165] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.161165] env[62235]: raise self.value [ 523.161165] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.161165] env[62235]: updated_port = self._update_port( [ 523.161165] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.161165] env[62235]: _ensure_no_port_binding_failure(port) [ 523.161165] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.161165] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.161945] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. [ 523.161945] env[62235]: Removing descriptor: 18 [ 523.161945] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Traceback (most recent call last): [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] yield resources [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self.driver.spawn(context, instance, image_meta, [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.161945] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] vm_ref = self.build_virtual_machine(instance, [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] for vif in network_info: [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return self._sync_wrapper(fn, *args, **kwargs) [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self.wait() [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self[:] = self._gt.wait() [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return self._exit_event.wait() [ 523.162515] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] result = hub.switch() [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return self.greenlet.switch() [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] result = function(*args, **kwargs) [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return func(*args, **kwargs) [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] raise e [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] nwinfo = self.network_api.allocate_for_instance( [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.162873] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] created_port_ids = self._update_ports_for_instance( [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] with excutils.save_and_reraise_exception(): [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self.force_reraise() [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] raise self.value [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] updated_port = self._update_port( [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] _ensure_no_port_binding_failure(port) [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.163223] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] raise exception.PortBindingFailed(port_id=port['id']) [ 523.163649] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] nova.exception.PortBindingFailed: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. [ 523.163649] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] [ 523.163649] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Terminating instance [ 523.165203] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.165355] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquired lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.165517] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.175216] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.188620] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271411, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49641} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.188918] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 523.189239] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 523.189462] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-16d5bdc3-2400-4e51-b673-62160b7bba57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.197899] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 523.197899] env[62235]: value = "task-1271412" [ 523.197899] env[62235]: _type = "Task" [ 523.197899] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.206957] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271412, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.208416] env[62235]: DEBUG nova.policy [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94bf8427bb65448ea2dc62d0dd05e519', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b19fc8b5c04d4311a1f5774aade4e0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.223739] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b11235-69ec-7398-16ae-3572140981c5, 'name': SearchDatastore_Task, 'duration_secs': 0.008465} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.224338] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.224338] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 4e20ae82-9a11-433b-ac4e-a85404cf4ec6/4e20ae82-9a11-433b-ac4e-a85404cf4ec6.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 523.224593] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aed17c63-6d06-4a44-bbcb-eabbdd2c6f46 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.234606] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 523.234606] env[62235]: value = "task-1271413" [ 523.234606] env[62235]: _type = "Task" [ 523.234606] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.245315] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.568869] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Acquiring lock "dbba80e1-52e2-4787-bcf7-c3300e923813" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.569198] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Lock "dbba80e1-52e2-4787-bcf7-c3300e923813" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.617812] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.681598] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Successfully created port: 84bdf075-27c4-4cd1-a686-e8e524c3585c {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.686067] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Releasing lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.686067] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 523.686067] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 523.686067] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86eeee38-3709-4003-a466-014b104ab4e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.700728] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9673829-cd6d-4afe-b92f-17a2997c90ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.731263] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271412, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072696} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.731601] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 523.732694] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65937237-96d7-4506-ab52-51f529530908 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.742739] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.745475] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eb6cc4b1-9236-417c-bdda-35afff032da5 could not be found. [ 523.745796] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 523.746292] env[62235]: INFO nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 523.746703] env[62235]: DEBUG oslo.service.loopingcall [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.752985] env[62235]: DEBUG nova.compute.manager [-] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 523.753119] env[62235]: DEBUG nova.network.neutron [-] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 523.781456] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 523.784664] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed37ee86-c55a-4f3a-8b70-4afa42e67cc6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.805559] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271413, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45164} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.806348] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 4e20ae82-9a11-433b-ac4e-a85404cf4ec6/4e20ae82-9a11-433b-ac4e-a85404cf4ec6.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 523.807326] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 523.807326] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85fa71ad-934a-4274-bc87-c4347ee6999f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.810737] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 523.810737] env[62235]: value = "task-1271414" [ 523.810737] env[62235]: _type = "Task" [ 523.810737] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.816120] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 523.816120] env[62235]: value = "task-1271415" [ 523.816120] env[62235]: _type = "Task" [ 523.816120] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.826362] env[62235]: DEBUG nova.network.neutron [-] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.827805] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271414, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.833715] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.029182] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8087e916-0304-495c-9f19-2793a50b96e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.040697] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a456d21-0199-4e01-9420-17d0e389bb9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.077884] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.079680] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15065d4-dd23-4013-b15a-b665543afd14 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.088119] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6fe801-8ba7-4e87-ae7b-1eaa098cdbf1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.109031] env[62235]: DEBUG nova.compute.provider_tree [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.332072] env[62235]: DEBUG nova.network.neutron [-] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.332432] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271414, 'name': ReconfigVM_Task, 'duration_secs': 0.307704} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.333163] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Reconfigured VM instance instance-00000002 to attach disk [datastore2] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 524.334313] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09a9b616-262b-47db-89a1-f0760be2e885 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.339628] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065338} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.343288] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 524.344155] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca64846-19b4-4148-b573-3608c7912d21 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.348704] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 524.348704] env[62235]: value = "task-1271416" [ 524.348704] env[62235]: _type = "Task" [ 524.348704] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.370521] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 4e20ae82-9a11-433b-ac4e-a85404cf4ec6/4e20ae82-9a11-433b-ac4e-a85404cf4ec6.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 524.372212] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b07a713-5ac2-4b2e-a624-af221ecab898 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.397330] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271416, 'name': Rename_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.398150] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 524.398150] env[62235]: value = "task-1271417" [ 524.398150] env[62235]: _type = "Task" [ 524.398150] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.407792] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271417, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.459276] env[62235]: DEBUG nova.compute.manager [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Received event network-changed-d9ab7d0d-d72b-40df-b207-fa6d04040e3f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 524.459276] env[62235]: DEBUG nova.compute.manager [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Refreshing instance network info cache due to event network-changed-d9ab7d0d-d72b-40df-b207-fa6d04040e3f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 524.459276] env[62235]: DEBUG oslo_concurrency.lockutils [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] Acquiring lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.459276] env[62235]: DEBUG oslo_concurrency.lockutils [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] Acquired lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.459276] env[62235]: DEBUG nova.network.neutron [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Refreshing network info cache for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 524.584694] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Releasing lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.585151] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.585345] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.585682] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc590c0f-9c98-44b0-98ef-41ed4b434caa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.598546] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78bbe387-eddb-4e99-a1c9-c8e39703a348 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.611392] env[62235]: DEBUG nova.scheduler.client.report [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.623659] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9617c521-d878-4749-a9b6-220ec51a2fc1 could not be found. [ 524.623941] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.624167] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 524.624445] env[62235]: DEBUG oslo.service.loopingcall [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.624986] env[62235]: DEBUG nova.compute.manager [-] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.625118] env[62235]: DEBUG nova.network.neutron [-] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.627889] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.661098] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.661098] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.661098] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.664700] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.664913] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.665083] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.665490] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.665490] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.665625] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.665781] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.666044] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.666870] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f324921a-5c6d-48f5-8274-a74a0c774285 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.676071] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7c6785-1ca1-44e5-8a8b-e1d261f12331 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.709054] env[62235]: DEBUG nova.network.neutron [-] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.835538] env[62235]: INFO nova.compute.manager [-] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Took 1.08 seconds to deallocate network for instance. [ 524.838232] env[62235]: DEBUG nova.compute.claims [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 524.838341] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.860815] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271416, 'name': Rename_Task, 'duration_secs': 0.132696} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.861018] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 524.861280] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61b4e34f-c639-4915-9eac-7918270716da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.867685] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 524.867685] env[62235]: value = "task-1271418" [ 524.867685] env[62235]: _type = "Task" [ 524.867685] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.875822] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271418, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.911972] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271417, 'name': ReconfigVM_Task, 'duration_secs': 0.263314} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.911972] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 4e20ae82-9a11-433b-ac4e-a85404cf4ec6/4e20ae82-9a11-433b-ac4e-a85404cf4ec6.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 524.913208] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52fde9bb-edb1-400c-a14e-0f8a978e3038 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.922164] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 524.922164] env[62235]: value = "task-1271419" [ 524.922164] env[62235]: _type = "Task" [ 524.922164] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.929472] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271419, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.954918] env[62235]: DEBUG nova.compute.manager [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Received event network-changed-0a437e0e-ed26-46ff-9867-052e8eca1eb2 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 524.955124] env[62235]: DEBUG nova.compute.manager [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Refreshing instance network info cache due to event network-changed-0a437e0e-ed26-46ff-9867-052e8eca1eb2. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 524.955332] env[62235]: DEBUG oslo_concurrency.lockutils [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] Acquiring lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.955472] env[62235]: DEBUG oslo_concurrency.lockutils [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] Acquired lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.955636] env[62235]: DEBUG nova.network.neutron [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Refreshing network info cache for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 524.993072] env[62235]: DEBUG nova.network.neutron [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.085646] env[62235]: DEBUG nova.network.neutron [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.126709] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.126995] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.133610] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.523s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.135467] env[62235]: INFO nova.compute.claims [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.215578] env[62235]: DEBUG nova.network.neutron [-] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.384772] env[62235]: DEBUG oslo_vmware.api [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271418, 'name': PowerOnVM_Task, 'duration_secs': 0.433547} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.384772] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 525.384772] env[62235]: INFO nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Took 10.02 seconds to spawn the instance on the hypervisor. [ 525.384968] env[62235]: DEBUG nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 525.385712] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1144c8ad-e7d4-4a4e-ad2c-bd2b874276ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.438969] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271419, 'name': Rename_Task, 'duration_secs': 0.143459} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.438969] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 525.438969] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2845f7d-30c3-4a58-a66f-173c3b6198c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.449132] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 525.449132] env[62235]: value = "task-1271420" [ 525.449132] env[62235]: _type = "Task" [ 525.449132] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.457458] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271420, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.507297] env[62235]: DEBUG nova.network.neutron [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.590024] env[62235]: DEBUG oslo_concurrency.lockutils [req-c626764c-0419-4c12-b796-bde704e2e3c1 req-ccbb380b-aa63-4220-825d-fb354685ef89 service nova] Releasing lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.643551] env[62235]: DEBUG nova.compute.utils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.651589] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.651768] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 525.720416] env[62235]: INFO nova.compute.manager [-] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Took 1.10 seconds to deallocate network for instance. [ 525.726073] env[62235]: DEBUG nova.compute.claims [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.726753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.761023] env[62235]: DEBUG nova.network.neutron [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.868275] env[62235]: DEBUG nova.policy [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94bf8427bb65448ea2dc62d0dd05e519', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b19fc8b5c04d4311a1f5774aade4e0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.912636] env[62235]: INFO nova.compute.manager [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Took 14.79 seconds to build instance. [ 525.965762] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271420, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.036344] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "7d0facf3-aa88-4332-8077-418db20be8b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.036607] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "7d0facf3-aa88-4332-8077-418db20be8b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.153239] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.264253] env[62235]: DEBUG oslo_concurrency.lockutils [req-9637a8f5-01c8-473c-b4ca-51578a35173b req-87815545-b620-4e07-a82d-3e3e050f8453 service nova] Releasing lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.406140] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2a86a8-f0a2-4cc0-a51c-45338a72a399 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.412374] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236a7a91-93ea-44f1-a518-693f2fce2d53 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.421133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3aa3bdb0-a96f-465c-a0bc-0f1f28686d0f tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.308s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.425022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 13.089s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.425022] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91339abd-da97-49f8-adda-e205fa3120f3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.480183] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce38873f-4533-40d9-afe2-30c3baff6ba7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.486789] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 526.486789] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 526.486789] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.486789] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 526.486789] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.486789] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 526.486789] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.486789] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 526.486789] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.486789] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 526.486789] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.486789] env[62235]: ERROR nova.compute.manager raise self.value [ 526.486789] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.486789] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 526.486789] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.486789] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 526.487280] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.487280] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 526.487280] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 526.487280] env[62235]: ERROR nova.compute.manager [ 526.487280] env[62235]: Traceback (most recent call last): [ 526.487280] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 526.487280] env[62235]: listener.cb(fileno) [ 526.487280] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.487280] env[62235]: result = function(*args, **kwargs) [ 526.487280] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.487280] env[62235]: return func(*args, **kwargs) [ 526.487280] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.487280] env[62235]: raise e [ 526.487280] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.487280] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 526.487280] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.487280] env[62235]: created_port_ids = self._update_ports_for_instance( [ 526.487280] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.487280] env[62235]: with excutils.save_and_reraise_exception(): [ 526.487280] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.487280] env[62235]: self.force_reraise() [ 526.487280] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.487280] env[62235]: raise self.value [ 526.487280] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.487280] env[62235]: updated_port = self._update_port( [ 526.487280] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.487280] env[62235]: _ensure_no_port_binding_failure(port) [ 526.487280] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.487280] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 526.488063] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 526.488063] env[62235]: Removing descriptor: 19 [ 526.489954] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] Traceback (most recent call last): [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] yield resources [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self.driver.spawn(context, instance, image_meta, [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] vm_ref = self.build_virtual_machine(instance, [ 526.489954] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] for vif in network_info: [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return self._sync_wrapper(fn, *args, **kwargs) [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self.wait() [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self[:] = self._gt.wait() [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return self._exit_event.wait() [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.490335] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] result = hub.switch() [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return self.greenlet.switch() [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] result = function(*args, **kwargs) [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return func(*args, **kwargs) [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] raise e [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] nwinfo = self.network_api.allocate_for_instance( [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] created_port_ids = self._update_ports_for_instance( [ 526.490654] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] with excutils.save_and_reraise_exception(): [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self.force_reraise() [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] raise self.value [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] updated_port = self._update_port( [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] _ensure_no_port_binding_failure(port) [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] raise exception.PortBindingFailed(port_id=port['id']) [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 526.490978] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] [ 526.491313] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Terminating instance [ 526.494105] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.494327] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquired lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.494502] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 526.503245] env[62235]: DEBUG oslo_vmware.api [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271420, 'name': PowerOnVM_Task, 'duration_secs': 0.600073} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.506163] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 526.506718] env[62235]: INFO nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Took 6.61 seconds to spawn the instance on the hypervisor. [ 526.506718] env[62235]: DEBUG nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 526.507848] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a539503a-6645-4dee-b7fc-18412d63cf8a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.519021] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faab077c-595e-46a5-b10a-7a0fbbe369dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.542154] env[62235]: DEBUG nova.compute.provider_tree [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.687012] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Successfully created port: 9aa6a36b-3b67-431a-a4fd-d6887a9d213e {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.925365] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.994263] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.571s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.037611] env[62235]: INFO nova.compute.manager [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Took 11.58 seconds to build instance. [ 527.045283] env[62235]: DEBUG nova.scheduler.client.report [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.077912] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.177781] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.207679] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.207903] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.208615] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.209064] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.209132] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.209342] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.209618] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.209891] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.210109] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.210217] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.210436] env[62235]: DEBUG nova.virt.hardware [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.220641] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813edec8-9c8c-4c3a-8b77-a3cf6f59ae67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.229497] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d7a05d-4b53-44a0-9bd6-dba9c7afb9ce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.333505] env[62235]: DEBUG nova.compute.manager [req-56fd1534-aa50-4138-aa27-81d8ade324d0 req-1b75a3fb-1ece-4d2d-8a53-867c187f0064 service nova] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Received event network-vif-deleted-d9ab7d0d-d72b-40df-b207-fa6d04040e3f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.361682] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.455912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.488596] env[62235]: INFO nova.compute.manager [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Rebuilding instance [ 527.541606] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37eaf4e5-7b91-4215-90f7-fb8d0bcfce93 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.102s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.552126] env[62235]: DEBUG nova.compute.manager [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 527.552126] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ac57cd-b942-4960-be95-8e5c77c4caa3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.555306] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.556729] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.559642] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.813s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.560984] env[62235]: INFO nova.compute.claims [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.575434] env[62235]: DEBUG nova.compute.manager [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Received event network-vif-deleted-0a437e0e-ed26-46ff-9867-052e8eca1eb2 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.575634] env[62235]: DEBUG nova.compute.manager [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] [instance: df5f7759-c872-4044-9901-200eb7933543] Received event network-changed-84bdf075-27c4-4cd1-a686-e8e524c3585c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.575789] env[62235]: DEBUG nova.compute.manager [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] [instance: df5f7759-c872-4044-9901-200eb7933543] Refreshing instance network info cache due to event network-changed-84bdf075-27c4-4cd1-a686-e8e524c3585c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 527.575966] env[62235]: DEBUG oslo_concurrency.lockutils [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] Acquiring lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.867455] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Releasing lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.868179] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 527.869695] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 527.869695] env[62235]: DEBUG oslo_concurrency.lockutils [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] Acquired lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.869695] env[62235]: DEBUG nova.network.neutron [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] [instance: df5f7759-c872-4044-9901-200eb7933543] Refreshing network info cache for port 84bdf075-27c4-4cd1-a686-e8e524c3585c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 527.870853] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18128d7e-2d7a-4f28-88a9-14005f6ce64e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.884030] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c225ec-9a4d-4032-8ed6-f2539af33eab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.918404] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df5f7759-c872-4044-9901-200eb7933543 could not be found. [ 527.918404] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 527.918404] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Took 0.05 seconds to destroy the instance on the hypervisor. [ 527.918404] env[62235]: DEBUG oslo.service.loopingcall [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.918404] env[62235]: DEBUG nova.compute.manager [-] [instance: df5f7759-c872-4044-9901-200eb7933543] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 527.918404] env[62235]: DEBUG nova.network.neutron [-] [instance: df5f7759-c872-4044-9901-200eb7933543] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 527.959042] env[62235]: DEBUG nova.network.neutron [-] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.046533] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.072753] env[62235]: DEBUG nova.compute.utils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.072753] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.072905] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 528.077168] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 528.078070] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1574f678-d7da-4a58-8700-0e498c4c8d37 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.086633] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 528.086633] env[62235]: value = "task-1271421" [ 528.086633] env[62235]: _type = "Task" [ 528.086633] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.101022] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271421, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.407595] env[62235]: DEBUG nova.network.neutron [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.463275] env[62235]: DEBUG nova.network.neutron [-] [instance: df5f7759-c872-4044-9901-200eb7933543] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.472657] env[62235]: DEBUG nova.policy [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef9461ccb2a74d12aa83f95c79e8b93d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8326d88e6cd74ebba28d6b4626b25c8d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.570697] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.577717] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.596238] env[62235]: DEBUG nova.network.neutron [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] [instance: df5f7759-c872-4044-9901-200eb7933543] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.600268] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271421, 'name': PowerOffVM_Task, 'duration_secs': 0.202089} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.600882] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 528.602922] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 528.604320] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7818dd84-52e0-462e-becd-98e41e504ae5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.623571] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 528.623571] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c16a86d3-8710-4edc-b6c8-71775dff8692 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.653748] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 528.653748] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 528.653748] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleting the datastore file [datastore2] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 528.659075] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7946c5ee-eb87-4d53-be1e-c8b3e6ff8e5e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.667588] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 528.667588] env[62235]: value = "task-1271423" [ 528.667588] env[62235]: _type = "Task" [ 528.667588] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.677079] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.750472] env[62235]: DEBUG nova.compute.manager [None req-ec1a9642-426c-4893-9192-be17cd471ed3 tempest-ServerDiagnosticsV248Test-1371903052 tempest-ServerDiagnosticsV248Test-1371903052-project-admin] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 528.752750] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1fe604-3397-4a8a-9e83-0ede8ee457a1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.770728] env[62235]: INFO nova.compute.manager [None req-ec1a9642-426c-4893-9192-be17cd471ed3 tempest-ServerDiagnosticsV248Test-1371903052 tempest-ServerDiagnosticsV248Test-1371903052-project-admin] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Retrieving diagnostics [ 528.771896] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22baefce-5d9d-4100-a78d-6cb31e2e9f52 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.878663] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce35ba5-7cee-4b5e-9fd4-4ae30a5d0c88 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.886410] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eb87ea-f0ba-4308-9239-b54e7f15b61c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.942718] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f35459-7149-4ba6-8c3f-faa1024c7cae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.952611] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca3c901-7d01-448d-9219-8e5819b2c12f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.972499] env[62235]: INFO nova.compute.manager [-] [instance: df5f7759-c872-4044-9901-200eb7933543] Took 1.06 seconds to deallocate network for instance. [ 528.972708] env[62235]: DEBUG nova.compute.provider_tree [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.979734] env[62235]: DEBUG nova.compute.claims [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 528.980221] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.101772] env[62235]: DEBUG oslo_concurrency.lockutils [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] Releasing lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.102078] env[62235]: DEBUG nova.compute.manager [req-f122957e-2858-4f70-b27e-a41b1a4dabb5 req-3edecd07-9dd6-43d7-8d02-6003d1a6c214 service nova] [instance: df5f7759-c872-4044-9901-200eb7933543] Received event network-vif-deleted-84bdf075-27c4-4cd1-a686-e8e524c3585c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.179514] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145015} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.180117] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 529.180365] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 529.180882] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 529.440180] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Acquiring lock "1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.440426] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Lock "1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.478827] env[62235]: DEBUG nova.scheduler.client.report [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.590861] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.627649] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.630695] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.630695] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.630695] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.630695] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.630695] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.631061] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.631061] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.631061] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.631061] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.631061] env[62235]: DEBUG nova.virt.hardware [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.631262] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ec351a-a832-4f6d-982d-8db0de9626f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.645081] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f3b2bc-7829-4c30-af91-3e60a734a751 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.858063] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Successfully created port: b4e8b95c-8686-4c0a-b650-19def10f8506 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.987134] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.987666] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.991821] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.252s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.993940] env[62235]: INFO nova.compute.claims [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.237991] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.242247] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.242247] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.242247] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.242247] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.242247] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.242853] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.242853] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.242853] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.242853] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.242853] env[62235]: DEBUG nova.virt.hardware [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.243068] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c4fb96-73d1-40d9-9d48-8ee85f626b57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.258186] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c42cd9a-97d2-4ca6-9606-385483d4da2e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.283124] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 530.289374] env[62235]: DEBUG oslo.service.loopingcall [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.289705] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 530.289989] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5778b6a-03e9-4688-979a-3d2e267d9ab1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.309152] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 530.309152] env[62235]: value = "task-1271424" [ 530.309152] env[62235]: _type = "Task" [ 530.309152] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.318336] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271424, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.339366] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.340098] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.340098] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 530.340098] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 530.496129] env[62235]: DEBUG nova.compute.utils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.497835] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.498018] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 530.602918] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 530.602918] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 530.602918] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.602918] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 530.602918] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.602918] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 530.602918] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.602918] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 530.602918] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.602918] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 530.602918] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.602918] env[62235]: ERROR nova.compute.manager raise self.value [ 530.602918] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.602918] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 530.602918] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.602918] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 530.606448] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.606448] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 530.606448] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 530.606448] env[62235]: ERROR nova.compute.manager [ 530.606448] env[62235]: Traceback (most recent call last): [ 530.606448] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 530.606448] env[62235]: listener.cb(fileno) [ 530.606448] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.606448] env[62235]: result = function(*args, **kwargs) [ 530.606448] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.606448] env[62235]: return func(*args, **kwargs) [ 530.606448] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.606448] env[62235]: raise e [ 530.606448] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.606448] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 530.606448] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.606448] env[62235]: created_port_ids = self._update_ports_for_instance( [ 530.606448] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.606448] env[62235]: with excutils.save_and_reraise_exception(): [ 530.606448] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.606448] env[62235]: self.force_reraise() [ 530.606448] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.606448] env[62235]: raise self.value [ 530.606448] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.606448] env[62235]: updated_port = self._update_port( [ 530.606448] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.606448] env[62235]: _ensure_no_port_binding_failure(port) [ 530.606448] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.606448] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 530.607437] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 530.607437] env[62235]: Removing descriptor: 18 [ 530.607437] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Traceback (most recent call last): [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] yield resources [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self.driver.spawn(context, instance, image_meta, [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self._vmops.spawn(context, instance, image_meta, injected_files, [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 530.607437] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] vm_ref = self.build_virtual_machine(instance, [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] vif_infos = vmwarevif.get_vif_info(self._session, [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] for vif in network_info: [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return self._sync_wrapper(fn, *args, **kwargs) [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self.wait() [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self[:] = self._gt.wait() [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return self._exit_event.wait() [ 530.607775] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] result = hub.switch() [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return self.greenlet.switch() [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] result = function(*args, **kwargs) [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return func(*args, **kwargs) [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] raise e [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] nwinfo = self.network_api.allocate_for_instance( [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.608385] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] created_port_ids = self._update_ports_for_instance( [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] with excutils.save_and_reraise_exception(): [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self.force_reraise() [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] raise self.value [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] updated_port = self._update_port( [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] _ensure_no_port_binding_failure(port) [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.608705] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] raise exception.PortBindingFailed(port_id=port['id']) [ 530.609772] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 530.609772] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] [ 530.609772] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Terminating instance [ 530.609772] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.609772] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquired lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.609772] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 530.713930] env[62235]: DEBUG nova.policy [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d7ef1bc26534632ad28710f9df9803c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e22686dbba4420dad98d5de1f4fd449', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.740988] env[62235]: DEBUG nova.compute.manager [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Received event network-changed-9aa6a36b-3b67-431a-a4fd-d6887a9d213e {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.741297] env[62235]: DEBUG nova.compute.manager [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Refreshing instance network info cache due to event network-changed-9aa6a36b-3b67-431a-a4fd-d6887a9d213e. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 530.742043] env[62235]: DEBUG oslo_concurrency.lockutils [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] Acquiring lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.828229] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271424, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.846028] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 530.846028] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 530.846028] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: df5f7759-c872-4044-9901-200eb7933543] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 530.846028] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 530.846028] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 530.846028] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 530.846827] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 530.846827] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.847151] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.847355] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 530.848669] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 531.001616] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.170584] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.266215] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d939a7-eaa4-477c-aacc-c7e608424c09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.274838] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a333ff89-07e2-4da9-ae12-9aed6798a4d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.310656] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd4ea84-e7b4-4965-be5b-8aae9e7c705f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.327938] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb21eb3-4189-4a7e-9303-28066010530b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.332551] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271424, 'name': CreateVM_Task, 'duration_secs': 0.520479} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.332913] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 531.334977] env[62235]: DEBUG oslo_vmware.service [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af855b12-6d03-44c9-a107-550b915478b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.346026] env[62235]: DEBUG nova.compute.provider_tree [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.350291] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.350447] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.350823] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 531.353835] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec0aa2e-1fe7-4194-a0b1-5d73dbdbf278 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.358359] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 531.358359] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524d26f6-c505-13d9-b081-6ad62c5bd014" [ 531.358359] env[62235]: _type = "Task" [ 531.358359] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.369743] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524d26f6-c505-13d9-b081-6ad62c5bd014, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.664797] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.849121] env[62235]: DEBUG nova.scheduler.client.report [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.875567] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.875814] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 531.876069] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.876210] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.876380] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 531.877231] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21fbd027-6a2c-4e0c-900a-c8177cd0ae3b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.902563] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 531.902563] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 531.902563] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23929a00-84ae-4c83-a496-e8516ce14bf0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.908989] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.916067] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4f39074-5f1b-4112-9a7d-ca34ebb906f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.921802] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 531.921802] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524101b7-68c0-34a6-a1bb-37c7995a2554" [ 531.921802] env[62235]: _type = "Task" [ 531.921802] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.931865] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524101b7-68c0-34a6-a1bb-37c7995a2554, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.014749] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.046878] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.047136] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.047300] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.047579] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.048039] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.048039] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.048192] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.048368] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.048645] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.051019] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.051019] env[62235]: DEBUG nova.virt.hardware [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.051019] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516497fa-635d-4071-a335-67e3b3f5c388 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.061126] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a8b30b-cc60-48bc-9d47-f0cf7ff9390c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.172288] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Releasing lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.173016] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 532.173016] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 532.173320] env[62235]: DEBUG oslo_concurrency.lockutils [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] Acquired lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.173393] env[62235]: DEBUG nova.network.neutron [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Refreshing network info cache for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 532.174653] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc2115c2-4add-495d-b2fb-052f96ee6305 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.185217] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0d3c44-4688-43de-abff-72e4b5b90219 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.216177] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dd37deb6-c5d8-44d5-8c2a-319ffce9f344 could not be found. [ 532.217191] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 532.217191] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Took 0.04 seconds to destroy the instance on the hypervisor. [ 532.217191] env[62235]: DEBUG oslo.service.loopingcall [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.217191] env[62235]: DEBUG nova.compute.manager [-] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.217191] env[62235]: DEBUG nova.network.neutron [-] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 532.259717] env[62235]: DEBUG nova.network.neutron [-] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.356641] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.356840] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.360747] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.114s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.364442] env[62235]: INFO nova.compute.claims [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.408431] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Successfully created port: 262d0fa9-cacb-4c38-8444-19c55535c54c {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.437126] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Preparing fetch location {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 532.437406] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating directory with path [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 532.437634] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e924ef4b-a3c8-479f-85f2-e6a99499302a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.470899] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Created directory with path [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 532.471133] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Fetch image to [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 532.471303] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Downloading image file data 02c64327-1eca-4d55-9f01-62fa8b2a2334 to [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk on the data store datastore1 {{(pid=62235) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 532.472140] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b78c841-0e45-4413-995e-937486e87017 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.482895] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ef2e0d-69c0-4180-98e2-d205910f1170 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.499406] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766f1666-a52c-414e-90cc-95eb6bbf4826 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.539240] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9999d3f1-5b60-4502-b7c7-2a0e5f85b1d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.543650] env[62235]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-855c2dfc-87ae-4611-a707-8e97ba877127 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.634195] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Downloading image file data 02c64327-1eca-4d55-9f01-62fa8b2a2334 to the data store datastore1 {{(pid=62235) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 532.682558] env[62235]: DEBUG oslo_vmware.rw_handles [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62235) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 532.751291] env[62235]: DEBUG nova.network.neutron [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.766411] env[62235]: DEBUG nova.network.neutron [-] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.790334] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.862629] env[62235]: DEBUG nova.compute.utils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.865255] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.865255] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 533.227940] env[62235]: DEBUG nova.policy [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d819ef90ee4aabbf4fe48e639beee6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c226dc4526f4455bf081061756b5727', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.268913] env[62235]: INFO nova.compute.manager [-] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Took 1.05 seconds to deallocate network for instance. [ 533.274418] env[62235]: DEBUG nova.compute.claims [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 533.274418] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.293790] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.294054] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 533.294744] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.295457] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.296121] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.296617] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.296889] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.300653] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.300885] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 533.301957] env[62235]: DEBUG nova.network.neutron [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.303514] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.335750] env[62235]: DEBUG oslo_vmware.rw_handles [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Completed reading data from the image iterator. {{(pid=62235) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 533.335750] env[62235]: DEBUG oslo_vmware.rw_handles [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 533.370272] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.477095] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Downloaded image file data 02c64327-1eca-4d55-9f01-62fa8b2a2334 to vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk on the data store datastore1 {{(pid=62235) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 533.479175] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Caching image {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 533.479562] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copying Virtual Disk [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk to [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 533.480134] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45307d47-aa42-4b04-aa44-22498db98f8f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.492163] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 533.492163] env[62235]: value = "task-1271425" [ 533.492163] env[62235]: _type = "Task" [ 533.492163] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.503590] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.588149] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866255c8-bc4a-41de-8432-9c30666cd27c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.599292] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa08f63-787a-48b6-9962-cb8648c44d54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.640571] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e445f047-b260-426b-b143-788a6d298b3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.648524] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fdf255-34ba-4420-bd1c-2b7e5501358c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.662761] env[62235]: DEBUG nova.compute.provider_tree [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.807351] env[62235]: DEBUG oslo_concurrency.lockutils [req-549571d1-4480-4a48-9bb0-028b57f769f3 req-a5729959-1564-4c56-856f-9391b9fedeb6 service nova] Releasing lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.811181] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.009568] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271425, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.166162] env[62235]: DEBUG nova.scheduler.client.report [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.378601] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.423704] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.424893] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.424893] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.424893] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.424893] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.424893] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.430021] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.430021] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.430021] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.430021] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.430021] env[62235]: DEBUG nova.virt.hardware [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.430292] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70f43d7-d79d-4608-97ef-efb8d082ef1b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.437322] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5edc12d-f74c-4441-96e8-8ed282f98ac4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.505943] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271425, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.746579} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.506313] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copied Virtual Disk [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk to [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 534.506537] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleting the datastore file [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334/tmp-sparse.vmdk {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 534.506819] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92755563-e4d3-4350-8e43-c64a10ca5c56 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.517176] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 534.517176] env[62235]: value = "task-1271426" [ 534.517176] env[62235]: _type = "Task" [ 534.517176] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.526552] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271426, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.638018] env[62235]: DEBUG nova.compute.manager [req-18127150-44ed-4daa-ae88-62028c6cb2ac req-1c54548b-8f85-474f-b18a-f3a00633ced4 service nova] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Received event network-vif-deleted-9aa6a36b-3b67-431a-a4fd-d6887a9d213e {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.671552] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.671552] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.676552] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.838s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.897907] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Acquiring lock "842f0697-6cec-423b-a35a-583985713a08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.898283] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Lock "842f0697-6cec-423b-a35a-583985713a08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.033830] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271426, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022679} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.034122] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 535.034335] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Moving file from [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd/02c64327-1eca-4d55-9f01-62fa8b2a2334 to [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334. {{(pid=62235) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 535.034591] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-14bcf304-7b53-4750-9caa-ae00478b3dbf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.044341] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 535.044341] env[62235]: value = "task-1271427" [ 535.044341] env[62235]: _type = "Task" [ 535.044341] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.059734] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271427, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.179302] env[62235]: DEBUG nova.compute.utils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.181649] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.182363] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 535.277084] env[62235]: ERROR nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. [ 535.277084] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 535.277084] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.277084] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 535.277084] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 535.277084] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 535.277084] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 535.277084] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 535.277084] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.277084] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 535.277084] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.277084] env[62235]: ERROR nova.compute.manager raise self.value [ 535.277084] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 535.277084] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 535.277084] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.277084] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 535.277845] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.277845] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 535.277845] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. [ 535.277845] env[62235]: ERROR nova.compute.manager [ 535.277845] env[62235]: Traceback (most recent call last): [ 535.277845] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 535.277845] env[62235]: listener.cb(fileno) [ 535.277845] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.277845] env[62235]: result = function(*args, **kwargs) [ 535.277845] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.277845] env[62235]: return func(*args, **kwargs) [ 535.277845] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.277845] env[62235]: raise e [ 535.277845] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.277845] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 535.277845] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 535.277845] env[62235]: created_port_ids = self._update_ports_for_instance( [ 535.277845] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 535.277845] env[62235]: with excutils.save_and_reraise_exception(): [ 535.277845] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.277845] env[62235]: self.force_reraise() [ 535.277845] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.277845] env[62235]: raise self.value [ 535.277845] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 535.277845] env[62235]: updated_port = self._update_port( [ 535.277845] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.277845] env[62235]: _ensure_no_port_binding_failure(port) [ 535.277845] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.277845] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 535.279868] env[62235]: nova.exception.PortBindingFailed: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. [ 535.279868] env[62235]: Removing descriptor: 19 [ 535.279868] env[62235]: ERROR nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Traceback (most recent call last): [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] yield resources [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self.driver.spawn(context, instance, image_meta, [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.279868] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] vm_ref = self.build_virtual_machine(instance, [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] for vif in network_info: [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return self._sync_wrapper(fn, *args, **kwargs) [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self.wait() [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self[:] = self._gt.wait() [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return self._exit_event.wait() [ 535.280984] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] result = hub.switch() [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return self.greenlet.switch() [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] result = function(*args, **kwargs) [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return func(*args, **kwargs) [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] raise e [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] nwinfo = self.network_api.allocate_for_instance( [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 535.281772] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] created_port_ids = self._update_ports_for_instance( [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] with excutils.save_and_reraise_exception(): [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self.force_reraise() [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] raise self.value [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] updated_port = self._update_port( [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] _ensure_no_port_binding_failure(port) [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.282497] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] raise exception.PortBindingFailed(port_id=port['id']) [ 535.283143] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] nova.exception.PortBindingFailed: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. [ 535.283143] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] [ 535.283143] env[62235]: INFO nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Terminating instance [ 535.284504] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Acquiring lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.284504] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Acquired lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.284504] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 535.374776] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Successfully created port: a66b3a5f-3385-48be-b820-39f6a3a54039 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.444791] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d967b8-6e85-408a-bf9f-0cca8b466cc2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.448588] env[62235]: DEBUG nova.policy [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e635bf5b9944ff5829118d947710c8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3d209198a13433cbb882e33f152460d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.457816] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6b833b-20c9-4afa-8265-5889684ad916 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.505974] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d31a23-5970-4182-97be-0ce4fd786849 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.516952] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee2e1b9-dc3a-4f3c-b669-b8e4f2e35bea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.530827] env[62235]: DEBUG nova.compute.provider_tree [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.557126] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271427, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025768} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.557451] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] File moved {{(pid=62235) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 535.558741] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Cleaning up location [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 535.558741] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleting the datastore file [datastore1] vmware_temp/eb2e2dde-20d4-4ef0-805c-1a99e8cf5ccd {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 535.558741] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4f57afa-891f-45cf-8846-cfe20ea671db {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.565015] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 535.565015] env[62235]: value = "task-1271428" [ 535.565015] env[62235]: _type = "Task" [ 535.565015] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.573321] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.688458] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.842126] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.035734] env[62235]: DEBUG nova.scheduler.client.report [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.078906] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0232} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.078906] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 536.083221] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.084723] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c730fab8-f4ca-4407-831f-ddd38d27fc7b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.097904] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 536.097904] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528a13e1-79cb-b735-7ab6-2a1fa452ca24" [ 536.097904] env[62235]: _type = "Task" [ 536.097904] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.112754] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528a13e1-79cb-b735-7ab6-2a1fa452ca24, 'name': SearchDatastore_Task, 'duration_secs': 0.008773} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.112754] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.112754] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 536.112754] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3f6f3d2-6820-4eca-ad32-0cefb4199925 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.123863] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 536.123863] env[62235]: value = "task-1271429" [ 536.123863] env[62235]: _type = "Task" [ 536.123863] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.136654] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "ab123e55-fa0e-4216-9afc-f8672598b08e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.136654] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "ab123e55-fa0e-4216-9afc-f8672598b08e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.145754] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271429, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.546552] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.547271] env[62235]: ERROR nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Traceback (most recent call last): [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self.driver.spawn(context, instance, image_meta, [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] vm_ref = self.build_virtual_machine(instance, [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.547271] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] for vif in network_info: [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return self._sync_wrapper(fn, *args, **kwargs) [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self.wait() [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self[:] = self._gt.wait() [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return self._exit_event.wait() [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] result = hub.switch() [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.547725] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return self.greenlet.switch() [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] result = function(*args, **kwargs) [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] return func(*args, **kwargs) [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] raise e [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] nwinfo = self.network_api.allocate_for_instance( [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] created_port_ids = self._update_ports_for_instance( [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] with excutils.save_and_reraise_exception(): [ 536.548105] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] self.force_reraise() [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] raise self.value [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] updated_port = self._update_port( [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] _ensure_no_port_binding_failure(port) [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] raise exception.PortBindingFailed(port_id=port['id']) [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] nova.exception.PortBindingFailed: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. [ 536.548533] env[62235]: ERROR nova.compute.manager [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] [ 536.551419] env[62235]: DEBUG nova.compute.utils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 536.551419] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.824s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.562319] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Build of instance eb6cc4b1-9236-417c-bdda-35afff032da5 was re-scheduled: Binding failed for port d9ab7d0d-d72b-40df-b207-fa6d04040e3f, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 536.562859] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 536.563138] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.563258] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquired lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.563413] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 536.593403] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Releasing lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.593833] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 536.594028] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 536.597384] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3e3d3aa-b6fe-493d-b59e-23789b546f9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.606316] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34398097-63bd-4040-8c7b-615618f6b7b3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.630504] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6d33eefc-b822-40fa-8b68-adb944e7dc7f could not be found. [ 536.630717] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 536.630895] env[62235]: INFO nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 536.631142] env[62235]: DEBUG oslo.service.loopingcall [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.634334] env[62235]: DEBUG nova.compute.manager [-] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.634430] env[62235]: DEBUG nova.network.neutron [-] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 536.641120] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271429, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450841} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.641360] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 536.641571] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 536.641831] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ffc3213-0587-4196-acd1-e8079aac0437 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.648046] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 536.648046] env[62235]: value = "task-1271430" [ 536.648046] env[62235]: _type = "Task" [ 536.648046] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.656073] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271430, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.700057] env[62235]: DEBUG nova.network.neutron [-] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.707610] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.734440] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.734585] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.734738] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.734921] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.735508] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.735981] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.737186] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.737339] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.737702] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.737702] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.737853] env[62235]: DEBUG nova.virt.hardware [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.738721] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10349b57-b98e-48cc-a896-e35aec4b89d0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.747125] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99eddf21-7922-4433-85bb-e286517391c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.033314] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Acquiring lock "b6122fc8-989d-4321-ac7a-80b2402a695f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.033614] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Lock "b6122fc8-989d-4321-ac7a-80b2402a695f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.130821] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.160780] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271430, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063289} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.161211] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 537.164523] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb2f496-5d40-44f5-909c-31399dceca22 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.186429] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 537.189280] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4b9c402-4860-4631-a14f-a40cee4aec9a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.210496] env[62235]: DEBUG nova.network.neutron [-] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.219168] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 537.219168] env[62235]: value = "task-1271431" [ 537.219168] env[62235]: _type = "Task" [ 537.219168] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.233440] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.278967] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Successfully created port: 7ba7ac15-7084-4fbe-ae35-80069dc46a3f {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.386554] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256541a7-fbfd-47fb-ba3a-6fd452ca4c3f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.394587] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b442c0e9-f616-44ff-9b52-f0793ee86370 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.429859] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fea55b9-dba7-449d-b205-9e24e5c61e12 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.438551] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d8c7ff-c993-4a67-abd2-716baf9c806e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.453174] env[62235]: DEBUG nova.compute.provider_tree [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.552175] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.713997] env[62235]: INFO nova.compute.manager [-] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Took 1.08 seconds to deallocate network for instance. [ 537.717719] env[62235]: DEBUG nova.compute.claims [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.717719] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.729103] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271431, 'name': ReconfigVM_Task, 'duration_secs': 0.259663} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.729333] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Reconfigured VM instance instance-00000002 to attach disk [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 537.729928] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98d0f2ea-adf6-4e0a-832a-81ede5f9b899 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.736608] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 537.736608] env[62235]: value = "task-1271432" [ 537.736608] env[62235]: _type = "Task" [ 537.736608] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.744894] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271432, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.957147] env[62235]: DEBUG nova.scheduler.client.report [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.055071] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Releasing lock "refresh_cache-eb6cc4b1-9236-417c-bdda-35afff032da5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.055324] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 538.055698] env[62235]: DEBUG nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.055978] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.098344] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.255249] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271432, 'name': Rename_Task, 'duration_secs': 0.133256} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.255742] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 538.255978] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44a67357-7b0f-45bd-988b-f92a59f122e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.263551] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 538.263551] env[62235]: value = "task-1271433" [ 538.263551] env[62235]: _type = "Task" [ 538.263551] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.271283] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.322175] env[62235]: ERROR nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. [ 538.322175] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.322175] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.322175] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.322175] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.322175] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.322175] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.322175] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.322175] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.322175] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 538.322175] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.322175] env[62235]: ERROR nova.compute.manager raise self.value [ 538.322175] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.322175] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.322175] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.322175] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.323102] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.323102] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.323102] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. [ 538.323102] env[62235]: ERROR nova.compute.manager [ 538.323102] env[62235]: Traceback (most recent call last): [ 538.323102] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.323102] env[62235]: listener.cb(fileno) [ 538.323102] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.323102] env[62235]: result = function(*args, **kwargs) [ 538.323102] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.323102] env[62235]: return func(*args, **kwargs) [ 538.323102] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.323102] env[62235]: raise e [ 538.323102] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.323102] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 538.323102] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.323102] env[62235]: created_port_ids = self._update_ports_for_instance( [ 538.323102] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.323102] env[62235]: with excutils.save_and_reraise_exception(): [ 538.323102] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.323102] env[62235]: self.force_reraise() [ 538.323102] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.323102] env[62235]: raise self.value [ 538.323102] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.323102] env[62235]: updated_port = self._update_port( [ 538.323102] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.323102] env[62235]: _ensure_no_port_binding_failure(port) [ 538.323102] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.323102] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.324082] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. [ 538.324082] env[62235]: Removing descriptor: 16 [ 538.324082] env[62235]: ERROR nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Traceback (most recent call last): [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] yield resources [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self.driver.spawn(context, instance, image_meta, [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.324082] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] vm_ref = self.build_virtual_machine(instance, [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] for vif in network_info: [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return self._sync_wrapper(fn, *args, **kwargs) [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self.wait() [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self[:] = self._gt.wait() [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return self._exit_event.wait() [ 538.324445] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] result = hub.switch() [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return self.greenlet.switch() [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] result = function(*args, **kwargs) [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return func(*args, **kwargs) [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] raise e [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] nwinfo = self.network_api.allocate_for_instance( [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.324809] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] created_port_ids = self._update_ports_for_instance( [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] with excutils.save_and_reraise_exception(): [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self.force_reraise() [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] raise self.value [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] updated_port = self._update_port( [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] _ensure_no_port_binding_failure(port) [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.325131] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] raise exception.PortBindingFailed(port_id=port['id']) [ 538.325437] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] nova.exception.PortBindingFailed: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. [ 538.325437] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] [ 538.325437] env[62235]: INFO nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Terminating instance [ 538.326732] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.326843] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.327010] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.462239] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.466010] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Traceback (most recent call last): [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self.driver.spawn(context, instance, image_meta, [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] vm_ref = self.build_virtual_machine(instance, [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.466010] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] for vif in network_info: [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return self._sync_wrapper(fn, *args, **kwargs) [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self.wait() [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self[:] = self._gt.wait() [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return self._exit_event.wait() [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] result = hub.switch() [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.466380] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return self.greenlet.switch() [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] result = function(*args, **kwargs) [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] return func(*args, **kwargs) [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] raise e [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] nwinfo = self.network_api.allocate_for_instance( [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] created_port_ids = self._update_ports_for_instance( [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] with excutils.save_and_reraise_exception(): [ 538.466737] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] self.force_reraise() [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] raise self.value [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] updated_port = self._update_port( [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] _ensure_no_port_binding_failure(port) [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] raise exception.PortBindingFailed(port_id=port['id']) [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] nova.exception.PortBindingFailed: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. [ 538.467082] env[62235]: ERROR nova.compute.manager [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] [ 538.468307] env[62235]: DEBUG nova.compute.utils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 538.470066] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.013s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.471152] env[62235]: INFO nova.compute.claims [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 538.475646] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Build of instance 9617c521-d878-4749-a9b6-220ec51a2fc1 was re-scheduled: Binding failed for port 0a437e0e-ed26-46ff-9867-052e8eca1eb2, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 538.479362] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 538.479640] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.480345] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquired lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.480345] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.603982] env[62235]: DEBUG nova.network.neutron [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.688243] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "c8efd1fa-0e33-423c-975e-b2583e1cfbd4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.688505] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "c8efd1fa-0e33-423c-975e-b2583e1cfbd4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.775119] env[62235]: DEBUG oslo_vmware.api [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271433, 'name': PowerOnVM_Task, 'duration_secs': 0.429454} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.775609] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 538.775609] env[62235]: DEBUG nova.compute.manager [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 538.776488] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff18433-dcc8-4777-98d6-65af84af3f6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.011230] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.045182] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.106719] env[62235]: INFO nova.compute.manager [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: eb6cc4b1-9236-417c-bdda-35afff032da5] Took 1.05 seconds to deallocate network for instance. [ 539.175351] env[62235]: DEBUG nova.compute.manager [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Received event network-changed-b4e8b95c-8686-4c0a-b650-19def10f8506 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.175542] env[62235]: DEBUG nova.compute.manager [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Refreshing instance network info cache due to event network-changed-b4e8b95c-8686-4c0a-b650-19def10f8506. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.175762] env[62235]: DEBUG oslo_concurrency.lockutils [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] Acquiring lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.175910] env[62235]: DEBUG oslo_concurrency.lockutils [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] Acquired lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.179318] env[62235]: DEBUG nova.network.neutron [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Refreshing network info cache for port b4e8b95c-8686-4c0a-b650-19def10f8506 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 539.206565] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.297854] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.324989] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.545741] env[62235]: DEBUG nova.compute.manager [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Received event network-changed-262d0fa9-cacb-4c38-8444-19c55535c54c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.545979] env[62235]: DEBUG nova.compute.manager [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Refreshing instance network info cache due to event network-changed-262d0fa9-cacb-4c38-8444-19c55535c54c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.546216] env[62235]: DEBUG oslo_concurrency.lockutils [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] Acquiring lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.709873] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Releasing lock "refresh_cache-9617c521-d878-4749-a9b6-220ec51a2fc1" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.710258] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 539.710823] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.710823] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.719867] env[62235]: ERROR nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. [ 539.719867] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.719867] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.719867] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.719867] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.719867] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.719867] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.719867] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.719867] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.719867] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 539.719867] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.719867] env[62235]: ERROR nova.compute.manager raise self.value [ 539.719867] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.719867] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.719867] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.719867] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.720386] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.720386] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.720386] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. [ 539.720386] env[62235]: ERROR nova.compute.manager [ 539.720386] env[62235]: Traceback (most recent call last): [ 539.720386] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.720386] env[62235]: listener.cb(fileno) [ 539.720386] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.720386] env[62235]: result = function(*args, **kwargs) [ 539.720386] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.720386] env[62235]: return func(*args, **kwargs) [ 539.720386] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.720386] env[62235]: raise e [ 539.720386] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.720386] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 539.720386] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.720386] env[62235]: created_port_ids = self._update_ports_for_instance( [ 539.720386] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.720386] env[62235]: with excutils.save_and_reraise_exception(): [ 539.720386] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.720386] env[62235]: self.force_reraise() [ 539.720386] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.720386] env[62235]: raise self.value [ 539.720386] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.720386] env[62235]: updated_port = self._update_port( [ 539.720386] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.720386] env[62235]: _ensure_no_port_binding_failure(port) [ 539.720386] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.720386] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.721135] env[62235]: nova.exception.PortBindingFailed: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. [ 539.721135] env[62235]: Removing descriptor: 18 [ 539.721135] env[62235]: ERROR nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Traceback (most recent call last): [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] yield resources [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self.driver.spawn(context, instance, image_meta, [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.721135] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] vm_ref = self.build_virtual_machine(instance, [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] for vif in network_info: [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return self._sync_wrapper(fn, *args, **kwargs) [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self.wait() [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self[:] = self._gt.wait() [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return self._exit_event.wait() [ 539.721498] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] result = hub.switch() [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return self.greenlet.switch() [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] result = function(*args, **kwargs) [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return func(*args, **kwargs) [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] raise e [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] nwinfo = self.network_api.allocate_for_instance( [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.721896] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] created_port_ids = self._update_ports_for_instance( [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] with excutils.save_and_reraise_exception(): [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self.force_reraise() [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] raise self.value [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] updated_port = self._update_port( [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] _ensure_no_port_binding_failure(port) [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.722237] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] raise exception.PortBindingFailed(port_id=port['id']) [ 539.722541] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] nova.exception.PortBindingFailed: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. [ 539.722541] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] [ 539.722541] env[62235]: INFO nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Terminating instance [ 539.724627] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.724627] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquired lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.724627] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 539.741382] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.749825] env[62235]: DEBUG nova.network.neutron [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.829837] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.830462] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.830844] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.831999] env[62235]: DEBUG oslo_concurrency.lockutils [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] Acquired lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.832094] env[62235]: DEBUG nova.network.neutron [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Refreshing network info cache for port 262d0fa9-cacb-4c38-8444-19c55535c54c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 539.835454] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a57e9fd-4d54-48d3-a115-9d178e43bbcc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.857723] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af1b008-dd63-4788-b4bc-ebb79d418c44 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.884958] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7 could not be found. [ 539.886130] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.886130] env[62235]: INFO nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Took 0.05 seconds to destroy the instance on the hypervisor. [ 539.886130] env[62235]: DEBUG oslo.service.loopingcall [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.886130] env[62235]: DEBUG nova.compute.manager [-] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.886130] env[62235]: DEBUG nova.network.neutron [-] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.904405] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f8495d-98ea-45a0-8a8e-be3c78c5f0ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.913140] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d583a34-7b69-4bf0-9d5c-e6686748107a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.951432] env[62235]: DEBUG nova.network.neutron [-] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.954223] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990736b3-deaa-4862-9734-65de2013e6cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.961016] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Acquiring lock "028e94f4-6b8f-45c3-a046-ed44e575362a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.961016] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Lock "028e94f4-6b8f-45c3-a046-ed44e575362a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.968756] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f0ebb9-cfa0-4a12-a3a5-b16675972d0c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.983720] env[62235]: DEBUG nova.compute.provider_tree [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.034632] env[62235]: DEBUG nova.network.neutron [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.149559] env[62235]: INFO nova.scheduler.client.report [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Deleted allocations for instance eb6cc4b1-9236-417c-bdda-35afff032da5 [ 540.247345] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.277517] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.364134] env[62235]: DEBUG nova.network.neutron [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.462031] env[62235]: DEBUG nova.network.neutron [-] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.490019] env[62235]: DEBUG nova.scheduler.client.report [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 540.507621] env[62235]: DEBUG nova.network.neutron [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.539615] env[62235]: DEBUG oslo_concurrency.lockutils [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] Releasing lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.540379] env[62235]: DEBUG nova.compute.manager [req-82e65fc7-7ff9-4986-a414-e3019956532d req-69b8565e-3b4d-4a7e-a8e3-9d2941946eb4 service nova] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Received event network-vif-deleted-b4e8b95c-8686-4c0a-b650-19def10f8506 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.557607] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.662744] env[62235]: DEBUG oslo_concurrency.lockutils [None req-be237125-42b5-470a-8d89-3124bd3191b7 tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "eb6cc4b1-9236-417c-bdda-35afff032da5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.565s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.757826] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: 9617c521-d878-4749-a9b6-220ec51a2fc1] Took 1.05 seconds to deallocate network for instance. [ 540.909335] env[62235]: DEBUG nova.compute.manager [None req-41ebc5d1-a8f7-4729-87c5-56ffd752d387 tempest-ServerDiagnosticsV248Test-1371903052 tempest-ServerDiagnosticsV248Test-1371903052-project-admin] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 540.910715] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79ec580-ae96-4f36-91db-2efaf9120858 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.919347] env[62235]: INFO nova.compute.manager [None req-41ebc5d1-a8f7-4729-87c5-56ffd752d387 tempest-ServerDiagnosticsV248Test-1371903052 tempest-ServerDiagnosticsV248Test-1371903052-project-admin] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Retrieving diagnostics [ 540.920340] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb36ded-5f1b-4a97-9a54-d81d9a18922e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.967297] env[62235]: INFO nova.compute.manager [-] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Took 1.08 seconds to deallocate network for instance. [ 540.972459] env[62235]: DEBUG nova.compute.claims [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.972818] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.995013] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.995626] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.998526] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.428s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.001084] env[62235]: INFO nova.compute.claims [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.011767] env[62235]: DEBUG oslo_concurrency.lockutils [req-1fc035eb-8e4e-495d-80cb-4cafd36947bc req-9c97da3e-f4f6-4ed3-b6e1-e45fce62ae50 service nova] Releasing lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.062317] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Releasing lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.062480] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.062690] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.069019] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ee2e4ef-33f7-4a77-ae45-c2830ec5198a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.075135] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Acquiring lock "41adaeb9-c39f-4214-b9d1-7137aab3455f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.075135] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Lock "41adaeb9-c39f-4214-b9d1-7137aab3455f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.082464] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63eaa256-102d-4ce8-ae33-12eba1bf714b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.113331] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d386f90d-25c3-4843-a91f-a092d27a6b5b could not be found. [ 541.113796] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.114046] env[62235]: INFO nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 541.114250] env[62235]: DEBUG oslo.service.loopingcall [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.114487] env[62235]: DEBUG nova.compute.manager [-] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.114583] env[62235]: DEBUG nova.network.neutron [-] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 541.165559] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.271890] env[62235]: DEBUG nova.network.neutron [-] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.468941] env[62235]: ERROR nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. [ 541.468941] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 541.468941] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.468941] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 541.468941] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.468941] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 541.468941] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.468941] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 541.468941] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.468941] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 541.468941] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.468941] env[62235]: ERROR nova.compute.manager raise self.value [ 541.468941] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.468941] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 541.468941] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.468941] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 541.469611] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.469611] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 541.469611] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. [ 541.469611] env[62235]: ERROR nova.compute.manager [ 541.469611] env[62235]: Traceback (most recent call last): [ 541.469611] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 541.469611] env[62235]: listener.cb(fileno) [ 541.469611] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.469611] env[62235]: result = function(*args, **kwargs) [ 541.469611] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.469611] env[62235]: return func(*args, **kwargs) [ 541.469611] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.469611] env[62235]: raise e [ 541.469611] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.469611] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 541.469611] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.469611] env[62235]: created_port_ids = self._update_ports_for_instance( [ 541.469611] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.469611] env[62235]: with excutils.save_and_reraise_exception(): [ 541.469611] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.469611] env[62235]: self.force_reraise() [ 541.469611] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.469611] env[62235]: raise self.value [ 541.469611] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.469611] env[62235]: updated_port = self._update_port( [ 541.469611] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.469611] env[62235]: _ensure_no_port_binding_failure(port) [ 541.469611] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.469611] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 541.470309] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. [ 541.470309] env[62235]: Removing descriptor: 22 [ 541.470309] env[62235]: ERROR nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Traceback (most recent call last): [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] yield resources [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self.driver.spawn(context, instance, image_meta, [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.470309] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] vm_ref = self.build_virtual_machine(instance, [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] for vif in network_info: [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return self._sync_wrapper(fn, *args, **kwargs) [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self.wait() [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self[:] = self._gt.wait() [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return self._exit_event.wait() [ 541.470617] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] result = hub.switch() [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return self.greenlet.switch() [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] result = function(*args, **kwargs) [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return func(*args, **kwargs) [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] raise e [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] nwinfo = self.network_api.allocate_for_instance( [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.470935] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] created_port_ids = self._update_ports_for_instance( [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] with excutils.save_and_reraise_exception(): [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self.force_reraise() [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] raise self.value [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] updated_port = self._update_port( [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] _ensure_no_port_binding_failure(port) [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.471259] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] raise exception.PortBindingFailed(port_id=port['id']) [ 541.471615] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] nova.exception.PortBindingFailed: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. [ 541.471615] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] [ 541.471615] env[62235]: INFO nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Terminating instance [ 541.475953] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Acquiring lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.475953] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Acquired lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.475953] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.505128] env[62235]: DEBUG nova.compute.utils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 541.513305] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 541.513850] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 541.584164] env[62235]: DEBUG nova.policy [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9de92e49eeda4c149aa67813ca3705fa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19fc2e8edbfd47e19bb77cf5bf1e91d7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 541.697779] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.775073] env[62235]: DEBUG nova.network.neutron [-] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.816595] env[62235]: INFO nova.scheduler.client.report [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Deleted allocations for instance 9617c521-d878-4749-a9b6-220ec51a2fc1 [ 542.010938] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.014688] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 542.160232] env[62235]: INFO nova.compute.manager [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Rebuilding instance [ 542.200208] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Successfully created port: 6acfaaad-2083-412d-8415-8cd96596e71d {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.218124] env[62235]: DEBUG nova.compute.manager [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 542.220795] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593db824-d2a4-4b3b-9154-7a833948af69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.252244] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.278125] env[62235]: INFO nova.compute.manager [-] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Took 1.16 seconds to deallocate network for instance. [ 542.281515] env[62235]: DEBUG nova.compute.claims [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.281755] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.328560] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "9617c521-d878-4749-a9b6-220ec51a2fc1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.109s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.349388] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e334f0de-4841-4360-9f2b-33a4de98b2c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.360215] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c6bd8c-07cc-4d67-9526-44fe4f644d43 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.398269] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39406589-eb0a-4553-a4ca-42366affe9e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.406899] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c69a65-136f-4386-b414-979abc2d84f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.423353] env[62235]: DEBUG nova.compute.provider_tree [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.510069] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Acquiring lock "e7684530-e42d-4a30-9c7c-28be8ff31b7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.510370] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Lock "e7684530-e42d-4a30-9c7c-28be8ff31b7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.513509] env[62235]: DEBUG nova.compute.manager [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Received event network-changed-a66b3a5f-3385-48be-b820-39f6a3a54039 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.513509] env[62235]: DEBUG nova.compute.manager [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Refreshing instance network info cache due to event network-changed-a66b3a5f-3385-48be-b820-39f6a3a54039. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.513509] env[62235]: DEBUG oslo_concurrency.lockutils [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] Acquiring lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.513509] env[62235]: DEBUG oslo_concurrency.lockutils [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] Acquired lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.513509] env[62235]: DEBUG nova.network.neutron [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Refreshing network info cache for port a66b3a5f-3385-48be-b820-39f6a3a54039 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 542.738766] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 542.738766] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a13b5bb6-599e-436d-97b9-526ac097d95b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.747014] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 542.747014] env[62235]: value = "task-1271434" [ 542.747014] env[62235]: _type = "Task" [ 542.747014] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.755175] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Releasing lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.758570] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.758570] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 542.760025] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23db6394-814f-4c1a-bb7c-ed788156d770 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.762174] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.777782] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74bce85-41cc-4721-ac41-191305329919 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.804675] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance db07ab9b-13a6-42e5-bc60-1a1919696f0e could not be found. [ 542.804887] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.805075] env[62235]: INFO nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 542.805334] env[62235]: DEBUG oslo.service.loopingcall [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.806028] env[62235]: DEBUG nova.compute.manager [-] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.806028] env[62235]: DEBUG nova.network.neutron [-] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.828565] env[62235]: DEBUG nova.network.neutron [-] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.833939] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.925155] env[62235]: DEBUG nova.compute.manager [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Received event network-vif-deleted-262d0fa9-cacb-4c38-8444-19c55535c54c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.925285] env[62235]: DEBUG nova.compute.manager [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Received event network-changed-7ba7ac15-7084-4fbe-ae35-80069dc46a3f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.925435] env[62235]: DEBUG nova.compute.manager [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Refreshing instance network info cache due to event network-changed-7ba7ac15-7084-4fbe-ae35-80069dc46a3f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.925724] env[62235]: DEBUG oslo_concurrency.lockutils [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] Acquiring lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.925873] env[62235]: DEBUG oslo_concurrency.lockutils [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] Acquired lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.926234] env[62235]: DEBUG nova.network.neutron [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Refreshing network info cache for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 542.928196] env[62235]: DEBUG nova.scheduler.client.report [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.029875] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 543.034958] env[62235]: DEBUG nova.network.neutron [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.061706] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.063320] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.064321] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.064562] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.064748] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.064959] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.065410] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.065715] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.065886] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.066813] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.066887] env[62235]: DEBUG nova.virt.hardware [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.068310] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f533d18-c9c9-4971-be5a-d53e0db1069a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.081020] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3114b39-9cba-4f23-b897-e474cd6d2912 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.128861] env[62235]: DEBUG nova.network.neutron [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.258032] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271434, 'name': PowerOffVM_Task, 'duration_secs': 0.126517} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.258032] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 543.258032] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 543.261240] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6968ae-7869-4af0-a72f-29fad5a64096 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.272042] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.272288] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.272493] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.272677] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.272844] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.283871] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 543.283871] env[62235]: INFO nova.compute.manager [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Terminating instance [ 543.284823] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5aaf2efe-23c4-43a3-aeca-ce23924cc40d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.296792] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "refresh_cache-4e20ae82-9a11-433b-ac4e-a85404cf4ec6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.296792] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquired lock "refresh_cache-4e20ae82-9a11-433b-ac4e-a85404cf4ec6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.296792] env[62235]: DEBUG nova.network.neutron [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.298271] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Acquiring lock "ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.298271] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Lock "ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.335479] env[62235]: DEBUG nova.network.neutron [-] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.368304] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.378909] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 543.379231] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 543.379231] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Deleting the datastore file [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 543.380502] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-183a2c30-edd6-4d28-ab59-cc440a971eee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.391463] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 543.391463] env[62235]: value = "task-1271436" [ 543.391463] env[62235]: _type = "Task" [ 543.391463] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.402719] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271436, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.438990] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.438990] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 543.440166] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.460s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.481350] env[62235]: DEBUG nova.network.neutron [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.633083] env[62235]: DEBUG oslo_concurrency.lockutils [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] Releasing lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.633083] env[62235]: DEBUG nova.compute.manager [req-3ef7f287-fb45-4dfa-a8cf-42d638e7d05b req-6b82100c-269a-4622-9be9-eb90c3aa3882 service nova] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Received event network-vif-deleted-a66b3a5f-3385-48be-b820-39f6a3a54039 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.634932] env[62235]: DEBUG nova.network.neutron [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.783317] env[62235]: ERROR nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. [ 543.783317] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.783317] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.783317] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.783317] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.783317] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.783317] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.783317] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.783317] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.783317] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 543.783317] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.783317] env[62235]: ERROR nova.compute.manager raise self.value [ 543.783317] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.783317] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.783317] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.783317] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.783868] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.783868] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.783868] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. [ 543.783868] env[62235]: ERROR nova.compute.manager [ 543.783868] env[62235]: Traceback (most recent call last): [ 543.783868] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.783868] env[62235]: listener.cb(fileno) [ 543.783868] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.783868] env[62235]: result = function(*args, **kwargs) [ 543.783868] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.783868] env[62235]: return func(*args, **kwargs) [ 543.783868] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.783868] env[62235]: raise e [ 543.783868] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.783868] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 543.783868] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.783868] env[62235]: created_port_ids = self._update_ports_for_instance( [ 543.783868] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.783868] env[62235]: with excutils.save_and_reraise_exception(): [ 543.783868] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.783868] env[62235]: self.force_reraise() [ 543.783868] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.783868] env[62235]: raise self.value [ 543.783868] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.783868] env[62235]: updated_port = self._update_port( [ 543.783868] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.783868] env[62235]: _ensure_no_port_binding_failure(port) [ 543.783868] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.783868] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.784820] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. [ 543.784820] env[62235]: Removing descriptor: 22 [ 543.784820] env[62235]: ERROR nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Traceback (most recent call last): [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] yield resources [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self.driver.spawn(context, instance, image_meta, [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.784820] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] vm_ref = self.build_virtual_machine(instance, [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] for vif in network_info: [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return self._sync_wrapper(fn, *args, **kwargs) [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self.wait() [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self[:] = self._gt.wait() [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return self._exit_event.wait() [ 543.787478] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] result = hub.switch() [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return self.greenlet.switch() [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] result = function(*args, **kwargs) [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return func(*args, **kwargs) [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] raise e [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] nwinfo = self.network_api.allocate_for_instance( [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.788311] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] created_port_ids = self._update_ports_for_instance( [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] with excutils.save_and_reraise_exception(): [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self.force_reraise() [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] raise self.value [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] updated_port = self._update_port( [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] _ensure_no_port_binding_failure(port) [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.790459] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] raise exception.PortBindingFailed(port_id=port['id']) [ 543.790809] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] nova.exception.PortBindingFailed: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. [ 543.790809] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] [ 543.790809] env[62235]: INFO nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Terminating instance [ 543.790809] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Acquiring lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.790809] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Acquired lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.790809] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.833994] env[62235]: DEBUG nova.network.neutron [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.843253] env[62235]: INFO nova.compute.manager [-] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Took 1.04 seconds to deallocate network for instance. [ 543.848523] env[62235]: DEBUG nova.compute.claims [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.848523] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.904087] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271436, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267714} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.904286] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 543.904809] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 543.904809] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 543.910578] env[62235]: DEBUG nova.network.neutron [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.945311] env[62235]: DEBUG nova.compute.utils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.949611] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 543.949877] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.011991] env[62235]: DEBUG nova.policy [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '488eb0da86a2488c8dbf2594fff89272', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f37ce761ef24a59902a405b8accd9d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.138517] env[62235]: DEBUG oslo_concurrency.lockutils [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] Releasing lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.138517] env[62235]: DEBUG nova.compute.manager [req-5d7ba6ab-d5da-4534-aede-3cbf3bc0e2f9 req-923e06c4-0b6b-44d7-ac67-38971e32a202 service nova] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Received event network-vif-deleted-7ba7ac15-7084-4fbe-ae35-80069dc46a3f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.299214] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000098f6-39bf-4bcf-988f-75c0862a277a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.308041] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960ea586-5fc3-43f9-bdcd-c2304349c27c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.314836] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.349549] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84423c82-4809-4498-9618-f9449cb4c205 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.357166] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742a051d-31bc-46ef-977d-edcdfad01b99 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.372298] env[62235]: DEBUG nova.compute.provider_tree [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.379437] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "a84e793b-4dbd-4b89-8fc4-94baacecd8e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.379437] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "a84e793b-4dbd-4b89-8fc4-94baacecd8e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.415351] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Releasing lock "refresh_cache-4e20ae82-9a11-433b-ac4e-a85404cf4ec6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.415831] env[62235]: DEBUG nova.compute.manager [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.416046] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.420242] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037a1481-e501-4bb6-99c1-0229ed170b51 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.429446] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "063af6b5-92ce-40b0-81e0-0dd6155898fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.429680] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "063af6b5-92ce-40b0-81e0-0dd6155898fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.435858] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 544.436377] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2950486d-56e3-412e-92c7-00dca659d163 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.445363] env[62235]: DEBUG oslo_vmware.api [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 544.445363] env[62235]: value = "task-1271437" [ 544.445363] env[62235]: _type = "Task" [ 544.445363] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.455303] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 544.458141] env[62235]: DEBUG oslo_vmware.api [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.497350] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Successfully created port: 4b47ecd1-f586-4dc0-9b2b-bc6926115db9 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.540611] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.600866] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Acquiring lock "7adac743-926e-4c66-8fbe-02b96fbdb2b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.600866] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Lock "7adac743-926e-4c66-8fbe-02b96fbdb2b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.880015] env[62235]: DEBUG nova.scheduler.client.report [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.952294] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.952575] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.952762] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.952941] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.953228] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.953308] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.954757] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.954757] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.954757] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.954757] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.954757] env[62235]: DEBUG nova.virt.hardware [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.960459] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d3c241-7dfe-403d-b0e3-dac09b6a755b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.977579] env[62235]: DEBUG oslo_vmware.api [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271437, 'name': PowerOffVM_Task, 'duration_secs': 0.126531} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.977665] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e645bda-c692-46ce-a71d-7443367557e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.982511] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 544.982738] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 544.983017] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-609f6485-d5a8-4ef9-9731-150ba802c8a6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.001370] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 545.008925] env[62235]: DEBUG oslo.service.loopingcall [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.009415] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 545.009782] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26809f66-a1e4-4bf0-a4e6-793a50afdfc6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.026601] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 545.027204] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 545.027522] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Deleting the datastore file [datastore2] 4e20ae82-9a11-433b-ac4e-a85404cf4ec6 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 545.028621] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-073b8a71-fd3f-49eb-b9b3-f7f554ed2fc5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.034364] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 545.034364] env[62235]: value = "task-1271439" [ 545.034364] env[62235]: _type = "Task" [ 545.034364] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.040518] env[62235]: DEBUG oslo_vmware.api [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for the task: (returnval){ [ 545.040518] env[62235]: value = "task-1271440" [ 545.040518] env[62235]: _type = "Task" [ 545.040518] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.047621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Releasing lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.048060] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 545.048219] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 545.048455] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271439, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.048989] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9d00d12-bbd6-430a-bd2c-344715a2584d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.056573] env[62235]: DEBUG oslo_vmware.api [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271440, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.061703] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da8f089-4da1-4747-8f76-10798aab19c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.091025] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dbba80e1-52e2-4787-bcf7-c3300e923813 could not be found. [ 545.091291] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 545.091542] env[62235]: INFO nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Took 0.04 seconds to destroy the instance on the hypervisor. [ 545.093016] env[62235]: DEBUG oslo.service.loopingcall [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.093016] env[62235]: DEBUG nova.compute.manager [-] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.093016] env[62235]: DEBUG nova.network.neutron [-] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.110547] env[62235]: DEBUG nova.network.neutron [-] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.393303] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.950s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.393303] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 545.393303] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] Traceback (most recent call last): [ 545.393303] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.393303] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self.driver.spawn(context, instance, image_meta, [ 545.393303] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.393303] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.393303] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.393303] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] vm_ref = self.build_virtual_machine(instance, [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] for vif in network_info: [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return self._sync_wrapper(fn, *args, **kwargs) [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self.wait() [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self[:] = self._gt.wait() [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return self._exit_event.wait() [ 545.393565] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] result = hub.switch() [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return self.greenlet.switch() [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] result = function(*args, **kwargs) [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] return func(*args, **kwargs) [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] raise e [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] nwinfo = self.network_api.allocate_for_instance( [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.393885] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] created_port_ids = self._update_ports_for_instance( [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] with excutils.save_and_reraise_exception(): [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] self.force_reraise() [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] raise self.value [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] updated_port = self._update_port( [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] _ensure_no_port_binding_failure(port) [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.394229] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] raise exception.PortBindingFailed(port_id=port['id']) [ 545.394539] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 545.394539] env[62235]: ERROR nova.compute.manager [instance: df5f7759-c872-4044-9901-200eb7933543] [ 545.394539] env[62235]: DEBUG nova.compute.utils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.394539] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.118s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.400671] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Build of instance df5f7759-c872-4044-9901-200eb7933543 was re-scheduled: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.401166] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.401687] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.401687] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquired lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.401824] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 545.468941] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 545.501669] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.501943] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.502075] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.502257] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.502397] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.502563] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.502766] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.502917] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.503124] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.503295] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.504162] env[62235]: DEBUG nova.virt.hardware [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.504347] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3db079-aa30-4416-9b14-7ad767cc0274 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.513614] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f565349-f2aa-4b9e-9e47-359674875e9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.544126] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271439, 'name': CreateVM_Task, 'duration_secs': 0.360375} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.550189] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 545.553483] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.554050] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.554050] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 545.554254] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Acquiring lock "74d069ed-c262-4b91-bcdd-ea3b4121adc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.557347] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Lock "74d069ed-c262-4b91-bcdd-ea3b4121adc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.557347] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9a86f06-16bd-479b-af17-f6d38fb745fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.562285] env[62235]: DEBUG oslo_vmware.api [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Task: {'id': task-1271440, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094462} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.562894] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 545.563075] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 545.564250] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 545.564250] env[62235]: INFO nova.compute.manager [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 545.564250] env[62235]: DEBUG oslo.service.loopingcall [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.564250] env[62235]: DEBUG nova.compute.manager [-] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.564250] env[62235]: DEBUG nova.network.neutron [-] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.567083] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 545.567083] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52778702-813a-e7e3-4540-7ba5efcc9ff1" [ 545.567083] env[62235]: _type = "Task" [ 545.567083] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.576031] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52778702-813a-e7e3-4540-7ba5efcc9ff1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.589546] env[62235]: DEBUG nova.network.neutron [-] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.613592] env[62235]: DEBUG nova.network.neutron [-] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.880119] env[62235]: ERROR nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. [ 545.880119] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.880119] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.880119] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.880119] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.880119] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.880119] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.880119] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.880119] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.880119] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 545.880119] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.880119] env[62235]: ERROR nova.compute.manager raise self.value [ 545.880119] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.880119] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.880119] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.880119] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.880827] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.880827] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.880827] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. [ 545.880827] env[62235]: ERROR nova.compute.manager [ 545.880827] env[62235]: Traceback (most recent call last): [ 545.880827] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.880827] env[62235]: listener.cb(fileno) [ 545.880827] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.880827] env[62235]: result = function(*args, **kwargs) [ 545.880827] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.880827] env[62235]: return func(*args, **kwargs) [ 545.880827] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.880827] env[62235]: raise e [ 545.880827] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.880827] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 545.880827] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.880827] env[62235]: created_port_ids = self._update_ports_for_instance( [ 545.880827] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.880827] env[62235]: with excutils.save_and_reraise_exception(): [ 545.880827] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.880827] env[62235]: self.force_reraise() [ 545.880827] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.880827] env[62235]: raise self.value [ 545.880827] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.880827] env[62235]: updated_port = self._update_port( [ 545.880827] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.880827] env[62235]: _ensure_no_port_binding_failure(port) [ 545.880827] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.880827] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.881584] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. [ 545.881584] env[62235]: Removing descriptor: 22 [ 545.881584] env[62235]: ERROR nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Traceback (most recent call last): [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] yield resources [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self.driver.spawn(context, instance, image_meta, [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.881584] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] vm_ref = self.build_virtual_machine(instance, [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] for vif in network_info: [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return self._sync_wrapper(fn, *args, **kwargs) [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self.wait() [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self[:] = self._gt.wait() [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return self._exit_event.wait() [ 545.881923] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] result = hub.switch() [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return self.greenlet.switch() [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] result = function(*args, **kwargs) [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return func(*args, **kwargs) [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] raise e [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] nwinfo = self.network_api.allocate_for_instance( [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.882272] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] created_port_ids = self._update_ports_for_instance( [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] with excutils.save_and_reraise_exception(): [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self.force_reraise() [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] raise self.value [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] updated_port = self._update_port( [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] _ensure_no_port_binding_failure(port) [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.882698] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] raise exception.PortBindingFailed(port_id=port['id']) [ 545.882997] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] nova.exception.PortBindingFailed: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. [ 545.882997] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] [ 545.882997] env[62235]: INFO nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Terminating instance [ 545.883545] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.883787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquired lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.883959] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 545.935039] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.078052] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52778702-813a-e7e3-4540-7ba5efcc9ff1, 'name': SearchDatastore_Task, 'duration_secs': 0.009201} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.080909] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.081324] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 546.081612] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.081773] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.081954] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 546.082551] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e913e24b-5c76-4b7c-821c-6febb91e1a76 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.089695] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 546.091382] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 546.091382] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fec5b2c8-1c58-4622-8cd8-0cf60c625bc9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.098364] env[62235]: DEBUG nova.network.neutron [-] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.102772] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 546.102772] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a76e0e-1324-d7bb-9031-5f60bcfa0fcc" [ 546.102772] env[62235]: _type = "Task" [ 546.102772] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.111208] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a76e0e-1324-d7bb-9031-5f60bcfa0fcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.115133] env[62235]: INFO nova.compute.manager [-] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Took 1.02 seconds to deallocate network for instance. [ 546.117698] env[62235]: DEBUG nova.compute.claims [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.118026] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.129723] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.266277] env[62235]: DEBUG nova.compute.manager [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Received event network-changed-6acfaaad-2083-412d-8415-8cd96596e71d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 546.266277] env[62235]: DEBUG nova.compute.manager [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Refreshing instance network info cache due to event network-changed-6acfaaad-2083-412d-8415-8cd96596e71d. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 546.266405] env[62235]: DEBUG oslo_concurrency.lockutils [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] Acquiring lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.266533] env[62235]: DEBUG oslo_concurrency.lockutils [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] Acquired lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.266662] env[62235]: DEBUG nova.network.neutron [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Refreshing network info cache for port 6acfaaad-2083-412d-8415-8cd96596e71d {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 546.279854] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e8eb8c-cf17-49c8-a529-c59c76c5f510 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.288889] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf1bc95-71ed-4b06-a92b-413404589e2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.326774] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a84b68-9439-4f01-b7f5-c8ac15fa6700 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.334880] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154ad496-e2c2-44d7-8141-f32612da3f57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.348793] env[62235]: DEBUG nova.compute.provider_tree [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.403272] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.558272] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.604020] env[62235]: INFO nova.compute.manager [-] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Took 1.04 seconds to deallocate network for instance. [ 546.618572] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a76e0e-1324-d7bb-9031-5f60bcfa0fcc, 'name': SearchDatastore_Task, 'duration_secs': 0.00897} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.619707] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae80cf7a-b728-4ed9-9fa2-2ad6dc910f01 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.627267] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 546.627267] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f91a39-388c-a7e6-7a2e-572ef55df629" [ 546.627267] env[62235]: _type = "Task" [ 546.627267] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.634473] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Releasing lock "refresh_cache-df5f7759-c872-4044-9901-200eb7933543" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.634591] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 546.635041] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.636593] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 546.636957] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f91a39-388c-a7e6-7a2e-572ef55df629, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.660851] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.790182] env[62235]: DEBUG nova.network.neutron [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.856221] env[62235]: DEBUG nova.scheduler.client.report [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.944483] env[62235]: DEBUG nova.network.neutron [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.060810] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Releasing lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.061283] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 547.062144] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.062144] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db6b35d3-a837-45e6-874a-bdd006b19735 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.075858] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d11331d-23b3-47a2-a71e-65e21cfed96f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.103716] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d0facf3-aa88-4332-8077-418db20be8b4 could not be found. [ 547.103957] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.104172] env[62235]: INFO nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 547.105427] env[62235]: DEBUG oslo.service.loopingcall [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.105427] env[62235]: DEBUG nova.compute.manager [-] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.105427] env[62235]: DEBUG nova.network.neutron [-] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 547.120251] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.128406] env[62235]: DEBUG nova.network.neutron [-] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.151825] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f91a39-388c-a7e6-7a2e-572ef55df629, 'name': SearchDatastore_Task, 'duration_secs': 0.009248} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.152624] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.152624] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 547.152624] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7793635c-b990-4703-bf34-b8ba384c63a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.160187] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 547.160187] env[62235]: value = "task-1271441" [ 547.160187] env[62235]: _type = "Task" [ 547.160187] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.164405] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.172381] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271441, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.361135] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.365179] env[62235]: ERROR nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Traceback (most recent call last): [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self.driver.spawn(context, instance, image_meta, [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] vm_ref = self.build_virtual_machine(instance, [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.365179] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] for vif in network_info: [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return self._sync_wrapper(fn, *args, **kwargs) [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self.wait() [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self[:] = self._gt.wait() [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return self._exit_event.wait() [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] result = hub.switch() [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.365555] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return self.greenlet.switch() [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] result = function(*args, **kwargs) [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] return func(*args, **kwargs) [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] raise e [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] nwinfo = self.network_api.allocate_for_instance( [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] created_port_ids = self._update_ports_for_instance( [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] with excutils.save_and_reraise_exception(): [ 547.366058] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] self.force_reraise() [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] raise self.value [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] updated_port = self._update_port( [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] _ensure_no_port_binding_failure(port) [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] raise exception.PortBindingFailed(port_id=port['id']) [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 547.366485] env[62235]: ERROR nova.compute.manager [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] [ 547.366889] env[62235]: DEBUG nova.compute.utils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.368572] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.556s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.368572] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.368572] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 547.368572] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.651s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.373369] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Build of instance dd37deb6-c5d8-44d5-8c2a-319ffce9f344 was re-scheduled: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.373369] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.373369] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquiring lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.373369] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Acquired lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.373558] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 547.374805] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1d8a65-24f1-486f-baa8-b81a3b117b3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.384059] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e85307c-40da-4740-909f-9972e6c425ff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.403562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0886a92-88e6-4603-adf1-02a5fde88433 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.411831] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7da39e-4bb4-4192-8531-a16272cfd7bd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.449513] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181514MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 547.449685] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.450318] env[62235]: DEBUG oslo_concurrency.lockutils [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] Releasing lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.450547] env[62235]: DEBUG nova.compute.manager [req-3ca03b86-4e2d-4e65-aa50-692438a562ef req-ab70192d-0c6f-4a9a-9d4b-9d0bcbd22a4d service nova] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Received event network-vif-deleted-6acfaaad-2083-412d-8415-8cd96596e71d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.639635] env[62235]: DEBUG nova.network.neutron [-] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.667546] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: df5f7759-c872-4044-9901-200eb7933543] Took 1.03 seconds to deallocate network for instance. [ 547.675154] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271441, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464533} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.675631] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 547.675834] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 547.676088] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cde131a5-d09c-4077-85a3-d865c9a9a4f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.683161] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 547.683161] env[62235]: value = "task-1271442" [ 547.683161] env[62235]: _type = "Task" [ 547.683161] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.690834] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.691095] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.704887] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.879408] env[62235]: DEBUG nova.compute.utils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Can not refresh info_cache because instance was not found {{(pid=62235) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 547.953693] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.144974] env[62235]: INFO nova.compute.manager [-] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Took 1.04 seconds to deallocate network for instance. [ 548.151908] env[62235]: DEBUG nova.compute.claims [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.152023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.199849] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069948} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.202220] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 548.203172] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b45cbe-a4be-424f-8d1c-3920975dacc9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.231098] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 548.232854] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65bd913c-cc8e-4a7d-893d-b888d515a337 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.256354] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 548.256354] env[62235]: value = "task-1271447" [ 548.256354] env[62235]: _type = "Task" [ 548.256354] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.269292] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271447, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.282413] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e632101-0d9b-4fb1-99d1-d22f1e11a6cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.290633] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.293371] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6484555-3dd7-45c3-a9c3-a954249a99a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.340746] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd63a33b-7f99-40db-af4a-3a9eef55061d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.350442] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0c95c2-a425-4897-87ef-f57747f709cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.373668] env[62235]: DEBUG nova.compute.provider_tree [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.379146] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "91e05f87-96a4-4d52-9016-841a3a034930" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.379146] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "91e05f87-96a4-4d52-9016-841a3a034930" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.769965] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271447, 'name': ReconfigVM_Task, 'duration_secs': 0.26948} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.769965] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Reconfigured VM instance instance-00000002 to attach disk [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0/d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 548.769965] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea458194-31ab-4864-afcd-1b2479a3911d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.778938] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 548.778938] env[62235]: value = "task-1271448" [ 548.778938] env[62235]: _type = "Task" [ 548.778938] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.787632] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271448, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.800594] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Releasing lock "refresh_cache-dd37deb6-c5d8-44d5-8c2a-319ffce9f344" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.800833] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 548.801020] env[62235]: DEBUG nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.801189] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.830044] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.838326] env[62235]: DEBUG nova.compute.manager [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Received event network-changed-4b47ecd1-f586-4dc0-9b2b-bc6926115db9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.838326] env[62235]: DEBUG nova.compute.manager [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Refreshing instance network info cache due to event network-changed-4b47ecd1-f586-4dc0-9b2b-bc6926115db9. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.838625] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] Acquiring lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.838625] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] Acquired lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.838783] env[62235]: DEBUG nova.network.neutron [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Refreshing network info cache for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 548.879752] env[62235]: DEBUG nova.scheduler.client.report [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.289603] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271448, 'name': Rename_Task, 'duration_secs': 0.139535} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.289869] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 549.290385] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6adbaee-6703-401d-9c8e-d3ea69a52af6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.296702] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Waiting for the task: (returnval){ [ 549.296702] env[62235]: value = "task-1271449" [ 549.296702] env[62235]: _type = "Task" [ 549.296702] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.304319] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.333381] env[62235]: DEBUG nova.network.neutron [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.376038] env[62235]: DEBUG nova.network.neutron [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.388194] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.388546] env[62235]: ERROR nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Traceback (most recent call last): [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self.driver.spawn(context, instance, image_meta, [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] vm_ref = self.build_virtual_machine(instance, [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.388546] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] for vif in network_info: [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return self._sync_wrapper(fn, *args, **kwargs) [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self.wait() [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self[:] = self._gt.wait() [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return self._exit_event.wait() [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] result = hub.switch() [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.389279] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return self.greenlet.switch() [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] result = function(*args, **kwargs) [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] return func(*args, **kwargs) [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] raise e [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] nwinfo = self.network_api.allocate_for_instance( [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] created_port_ids = self._update_ports_for_instance( [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] with excutils.save_and_reraise_exception(): [ 549.389943] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] self.force_reraise() [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] raise self.value [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] updated_port = self._update_port( [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] _ensure_no_port_binding_failure(port) [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] raise exception.PortBindingFailed(port_id=port['id']) [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] nova.exception.PortBindingFailed: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. [ 549.390710] env[62235]: ERROR nova.compute.manager [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] [ 549.391120] env[62235]: DEBUG nova.compute.utils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 549.394994] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.097s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.395201] env[62235]: DEBUG nova.objects.instance [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 549.399401] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Build of instance 6d33eefc-b822-40fa-8b68-adb944e7dc7f was re-scheduled: Binding failed for port b4e8b95c-8686-4c0a-b650-19def10f8506, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 549.399505] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 549.399737] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Acquiring lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.399890] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Acquired lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.400784] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.555876] env[62235]: DEBUG nova.network.neutron [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.719880] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "df5f7759-c872-4044-9901-200eb7933543" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.448s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.719880] env[62235]: Traceback (most recent call last): [ 549.719880] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.719880] env[62235]: self.driver.spawn(context, instance, image_meta, [ 549.719880] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 549.719880] env[62235]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.719880] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.719880] env[62235]: vm_ref = self.build_virtual_machine(instance, [ 549.719880] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.719880] env[62235]: vif_infos = vmwarevif.get_vif_info(self._session, [ 549.719880] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.719880] env[62235]: for vif in network_info: [ 549.719880] env[62235]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.719880] env[62235]: return self._sync_wrapper(fn, *args, **kwargs) [ 549.719880] env[62235]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.719880] env[62235]: self.wait() [ 549.719880] env[62235]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.719880] env[62235]: self[:] = self._gt.wait() [ 549.719880] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.719880] env[62235]: return self._exit_event.wait() [ 549.719880] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.719880] env[62235]: result = hub.switch() [ 549.719880] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.719880] env[62235]: return self.greenlet.switch() [ 549.719880] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.719880] env[62235]: result = function(*args, **kwargs) [ 549.720738] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.720738] env[62235]: return func(*args, **kwargs) [ 549.720738] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.720738] env[62235]: raise e [ 549.720738] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.720738] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 549.720738] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.720738] env[62235]: created_port_ids = self._update_ports_for_instance( [ 549.720738] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.720738] env[62235]: with excutils.save_and_reraise_exception(): [ 549.720738] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.720738] env[62235]: self.force_reraise() [ 549.720738] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.720738] env[62235]: raise self.value [ 549.720738] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.720738] env[62235]: updated_port = self._update_port( [ 549.720738] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.720738] env[62235]: _ensure_no_port_binding_failure(port) [ 549.720738] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.720738] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.720738] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 549.720738] env[62235]: During handling of the above exception, another exception occurred: [ 549.720738] env[62235]: Traceback (most recent call last): [ 549.720738] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 549.720738] env[62235]: self._build_and_run_instance(context, instance, image, [ 549.720738] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 549.720738] env[62235]: raise exception.RescheduledException( [ 549.720738] env[62235]: nova.exception.RescheduledException: Build of instance df5f7759-c872-4044-9901-200eb7933543 was re-scheduled: Binding failed for port 84bdf075-27c4-4cd1-a686-e8e524c3585c, please check neutron logs for more information. [ 549.721547] env[62235]: During handling of the above exception, another exception occurred: [ 549.721547] env[62235]: Traceback (most recent call last): [ 549.721547] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 549.721547] env[62235]: func(*args, **kwargs) [ 549.721547] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.721547] env[62235]: return func(*args, **kwargs) [ 549.721547] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 549.721547] env[62235]: return f(*args, **kwargs) [ 549.721547] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 549.721547] env[62235]: result = self._do_build_and_run_instance(*args, **kwargs) [ 549.721547] env[62235]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 549.721547] env[62235]: with excutils.save_and_reraise_exception(): [ 549.721547] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.721547] env[62235]: self.force_reraise() [ 549.721547] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.721547] env[62235]: raise self.value [ 549.721547] env[62235]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 549.721547] env[62235]: return f(self, context, *args, **kw) [ 549.721547] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 549.721547] env[62235]: with excutils.save_and_reraise_exception(): [ 549.721547] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.721547] env[62235]: self.force_reraise() [ 549.721547] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.721547] env[62235]: raise self.value [ 549.721547] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 549.721547] env[62235]: return function(self, context, *args, **kwargs) [ 549.721547] env[62235]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 549.721547] env[62235]: return function(self, context, *args, **kwargs) [ 549.721547] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 549.721547] env[62235]: return function(self, context, *args, **kwargs) [ 549.721547] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 549.721547] env[62235]: instance.save() [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 549.722373] env[62235]: updates, result = self.indirection_api.object_action( [ 549.722373] env[62235]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 549.722373] env[62235]: return cctxt.call(context, 'object_action', objinst=objinst, [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 549.722373] env[62235]: result = self.transport._send( [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 549.722373] env[62235]: return self._driver.send(target, ctxt, message, [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 549.722373] env[62235]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 549.722373] env[62235]: raise result [ 549.722373] env[62235]: nova.exception_Remote.InstanceNotFound_Remote: Instance df5f7759-c872-4044-9901-200eb7933543 could not be found. [ 549.722373] env[62235]: Traceback (most recent call last): [ 549.722373] env[62235]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 549.722373] env[62235]: return getattr(target, method)(*args, **kwargs) [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 549.722373] env[62235]: return fn(self, *args, **kwargs) [ 549.722373] env[62235]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 549.722373] env[62235]: old_ref, inst_ref = db.instance_update_and_get_original( [ 549.722373] env[62235]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 549.722373] env[62235]: return f(*args, **kwargs) [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 549.722373] env[62235]: with excutils.save_and_reraise_exception() as ectxt: [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.722373] env[62235]: self.force_reraise() [ 549.722373] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.722373] env[62235]: raise self.value [ 549.724709] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 549.724709] env[62235]: return f(*args, **kwargs) [ 549.724709] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 549.724709] env[62235]: return f(context, *args, **kwargs) [ 549.724709] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 549.724709] env[62235]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 549.724709] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 549.724709] env[62235]: raise exception.InstanceNotFound(instance_id=uuid) [ 549.724709] env[62235]: nova.exception.InstanceNotFound: Instance df5f7759-c872-4044-9901-200eb7933543 could not be found. [ 549.812436] env[62235]: DEBUG oslo_vmware.api [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Task: {'id': task-1271449, 'name': PowerOnVM_Task, 'duration_secs': 0.442629} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.812436] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 549.813237] env[62235]: DEBUG nova.compute.manager [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 549.814148] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1bacbe-4017-49fc-88a9-4377dd43a00a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.836073] env[62235]: INFO nova.compute.manager [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] [instance: dd37deb6-c5d8-44d5-8c2a-319ffce9f344] Took 1.03 seconds to deallocate network for instance. [ 550.059065] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] Releasing lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.059479] env[62235]: DEBUG nova.compute.manager [req-7e4b6f37-9b46-4dc6-b6ba-eb9cda66b43f req-0fa45257-60c6-4913-ace6-4a6399433fb2 service nova] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Received event network-vif-deleted-4b47ecd1-f586-4dc0-9b2b-bc6926115db9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.059909] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "279097d2-348d-4f58-88ec-8e2e3143f754" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.060246] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "279097d2-348d-4f58-88ec-8e2e3143f754" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.102676] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.222697] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.273517] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.331648] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.415023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f86f5ef3-f24c-47f6-93ec-d82d564b4184 tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.415023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.441s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.754037] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.777293] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Releasing lock "refresh_cache-6d33eefc-b822-40fa-8b68-adb944e7dc7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.777556] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 550.777728] env[62235]: DEBUG nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.777922] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.839396] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.345951] env[62235]: DEBUG nova.network.neutron [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.363364] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19cd027d-f078-4cba-9e7d-3f7e0b03f076 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.374894] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0d9dfd-d40c-4a09-8ee8-fe3aa455fb06 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.412091] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80026be8-8a48-43a5-a638-4423590765ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.420811] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7617cd-fee3-4c06-bbe8-5c47819175ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.441321] env[62235]: DEBUG nova.compute.provider_tree [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.850428] env[62235]: INFO nova.compute.manager [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] [instance: 6d33eefc-b822-40fa-8b68-adb944e7dc7f] Took 1.07 seconds to deallocate network for instance. [ 551.890269] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7691cb4b-4ca6-4e9f-8456-f1ac6bcd4fed tempest-ListServersNegativeTestJSON-475135279 tempest-ListServersNegativeTestJSON-475135279-project-member] Lock "dd37deb6-c5d8-44d5-8c2a-319ffce9f344" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.586s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.895226] env[62235]: Traceback (most recent call last): [ 551.895226] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.895226] env[62235]: self.driver.spawn(context, instance, image_meta, [ 551.895226] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.895226] env[62235]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.895226] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.895226] env[62235]: vm_ref = self.build_virtual_machine(instance, [ 551.895226] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.895226] env[62235]: vif_infos = vmwarevif.get_vif_info(self._session, [ 551.895226] env[62235]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.895226] env[62235]: for vif in network_info: [ 551.895226] env[62235]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.895226] env[62235]: return self._sync_wrapper(fn, *args, **kwargs) [ 551.895226] env[62235]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.895226] env[62235]: self.wait() [ 551.895226] env[62235]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.895226] env[62235]: self[:] = self._gt.wait() [ 551.895226] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.895226] env[62235]: return self._exit_event.wait() [ 551.895226] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.895226] env[62235]: result = hub.switch() [ 551.895226] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.895226] env[62235]: return self.greenlet.switch() [ 551.895226] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.895226] env[62235]: result = function(*args, **kwargs) [ 551.895226] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.895226] env[62235]: return func(*args, **kwargs) [ 551.895226] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.895226] env[62235]: raise e [ 551.895226] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.895226] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 551.895226] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.895226] env[62235]: created_port_ids = self._update_ports_for_instance( [ 551.896270] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.896270] env[62235]: with excutils.save_and_reraise_exception(): [ 551.896270] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.896270] env[62235]: self.force_reraise() [ 551.896270] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.896270] env[62235]: raise self.value [ 551.896270] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.896270] env[62235]: updated_port = self._update_port( [ 551.896270] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.896270] env[62235]: _ensure_no_port_binding_failure(port) [ 551.896270] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.896270] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.896270] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 551.896270] env[62235]: During handling of the above exception, another exception occurred: [ 551.896270] env[62235]: Traceback (most recent call last): [ 551.896270] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 551.896270] env[62235]: self._build_and_run_instance(context, instance, image, [ 551.896270] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 551.896270] env[62235]: raise exception.RescheduledException( [ 551.896270] env[62235]: nova.exception.RescheduledException: Build of instance dd37deb6-c5d8-44d5-8c2a-319ffce9f344 was re-scheduled: Binding failed for port 9aa6a36b-3b67-431a-a4fd-d6887a9d213e, please check neutron logs for more information. [ 551.896270] env[62235]: During handling of the above exception, another exception occurred: [ 551.896270] env[62235]: Traceback (most recent call last): [ 551.896270] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 551.896270] env[62235]: func(*args, **kwargs) [ 551.896270] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.896270] env[62235]: return func(*args, **kwargs) [ 551.896270] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 551.896270] env[62235]: return f(*args, **kwargs) [ 551.898409] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 551.898409] env[62235]: result = self._do_build_and_run_instance(*args, **kwargs) [ 551.898409] env[62235]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 551.898409] env[62235]: with excutils.save_and_reraise_exception(): [ 551.898409] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.898409] env[62235]: self.force_reraise() [ 551.898409] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.898409] env[62235]: raise self.value [ 551.898409] env[62235]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 551.898409] env[62235]: return f(self, context, *args, **kw) [ 551.898409] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 551.898409] env[62235]: with excutils.save_and_reraise_exception(): [ 551.898409] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.898409] env[62235]: self.force_reraise() [ 551.898409] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.898409] env[62235]: raise self.value [ 551.898409] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 551.898409] env[62235]: return function(self, context, *args, **kwargs) [ 551.898409] env[62235]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 551.898409] env[62235]: return function(self, context, *args, **kwargs) [ 551.898409] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 551.898409] env[62235]: return function(self, context, *args, **kwargs) [ 551.898409] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 551.898409] env[62235]: instance.save() [ 551.898409] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 551.898409] env[62235]: updates, result = self.indirection_api.object_action( [ 551.898409] env[62235]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 551.898409] env[62235]: return cctxt.call(context, 'object_action', objinst=objinst, [ 551.898409] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 551.898409] env[62235]: result = self.transport._send( [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 551.899258] env[62235]: return self._driver.send(target, ctxt, message, [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 551.899258] env[62235]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 551.899258] env[62235]: raise result [ 551.899258] env[62235]: nova.exception_Remote.InstanceNotFound_Remote: Instance dd37deb6-c5d8-44d5-8c2a-319ffce9f344 could not be found. [ 551.899258] env[62235]: Traceback (most recent call last): [ 551.899258] env[62235]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 551.899258] env[62235]: return getattr(target, method)(*args, **kwargs) [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 551.899258] env[62235]: return fn(self, *args, **kwargs) [ 551.899258] env[62235]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 551.899258] env[62235]: old_ref, inst_ref = db.instance_update_and_get_original( [ 551.899258] env[62235]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 551.899258] env[62235]: return f(*args, **kwargs) [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 551.899258] env[62235]: with excutils.save_and_reraise_exception() as ectxt: [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.899258] env[62235]: self.force_reraise() [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.899258] env[62235]: raise self.value [ 551.899258] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 551.899258] env[62235]: return f(*args, **kwargs) [ 551.899258] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 551.899258] env[62235]: return f(context, *args, **kwargs) [ 551.899258] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 551.899258] env[62235]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 551.899258] env[62235]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 551.900576] env[62235]: raise exception.InstanceNotFound(instance_id=uuid) [ 551.900576] env[62235]: nova.exception.InstanceNotFound: Instance dd37deb6-c5d8-44d5-8c2a-319ffce9f344 could not be found. [ 551.943611] env[62235]: DEBUG nova.scheduler.client.report [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.274381] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.274381] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.274381] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.274381] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.274645] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.279177] env[62235]: INFO nova.compute.manager [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Terminating instance [ 552.283902] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "refresh_cache-d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.284916] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquired lock "refresh_cache-d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.284916] env[62235]: DEBUG nova.network.neutron [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.401907] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.451350] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.451350] env[62235]: ERROR nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. [ 552.451350] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Traceback (most recent call last): [ 552.451350] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.451350] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self.driver.spawn(context, instance, image_meta, [ 552.451350] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 552.451350] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.451350] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.451350] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] vm_ref = self.build_virtual_machine(instance, [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] for vif in network_info: [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return self._sync_wrapper(fn, *args, **kwargs) [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self.wait() [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self[:] = self._gt.wait() [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return self._exit_event.wait() [ 552.451731] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] result = hub.switch() [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return self.greenlet.switch() [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] result = function(*args, **kwargs) [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] return func(*args, **kwargs) [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] raise e [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] nwinfo = self.network_api.allocate_for_instance( [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.452076] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] created_port_ids = self._update_ports_for_instance( [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] with excutils.save_and_reraise_exception(): [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] self.force_reraise() [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] raise self.value [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] updated_port = self._update_port( [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] _ensure_no_port_binding_failure(port) [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.452578] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] raise exception.PortBindingFailed(port_id=port['id']) [ 552.452923] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] nova.exception.PortBindingFailed: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. [ 552.452923] env[62235]: ERROR nova.compute.manager [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] [ 552.452923] env[62235]: DEBUG nova.compute.utils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 552.453257] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Build of instance 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7 was re-scheduled: Binding failed for port 262d0fa9-cacb-4c38-8444-19c55535c54c, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 552.455022] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 552.455482] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.455796] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.456278] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.457821] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.761s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.459576] env[62235]: INFO nova.compute.claims [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.872535] env[62235]: DEBUG nova.network.neutron [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.909256] env[62235]: INFO nova.scheduler.client.report [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Deleted allocations for instance 6d33eefc-b822-40fa-8b68-adb944e7dc7f [ 552.945185] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.211192] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.394993] env[62235]: DEBUG nova.network.neutron [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.420058] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ac95c847-801c-43de-92e2-2b9f2bb7cedb tempest-ServerExternalEventsTest-1519852148 tempest-ServerExternalEventsTest-1519852148-project-member] Lock "6d33eefc-b822-40fa-8b68-adb944e7dc7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.858s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.621379] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.885691] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b72b617-1760-43ec-b685-f37946bf89b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.894475] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be7c0c5-d3ba-4403-872d-6f3a6a8eb2e0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.898594] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Releasing lock "refresh_cache-d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.899014] env[62235]: DEBUG nova.compute.manager [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 553.899208] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.901087] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520118c5-867d-4664-a5ec-04b32ebe7a11 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.908645] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 553.938921] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4d00268-44cf-4959-97a7-a598185c5ee9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.941228] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.944805] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb00ef9-5079-4d3b-bb29-fbdbbf8f4256 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.957473] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec881675-3c7e-4e30-a790-5c3759830fa0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.961954] env[62235]: DEBUG oslo_vmware.api [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 553.961954] env[62235]: value = "task-1271452" [ 553.961954] env[62235]: _type = "Task" [ 553.961954] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.977247] env[62235]: DEBUG nova.compute.provider_tree [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.987136] env[62235]: DEBUG oslo_vmware.api [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.124672] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "refresh_cache-3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.126025] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.128016] env[62235]: DEBUG nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.128016] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.179853] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.471997] env[62235]: DEBUG oslo_vmware.api [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271452, 'name': PowerOffVM_Task, 'duration_secs': 0.122113} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.473053] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.473053] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 554.473180] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 554.473822] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2036e27c-58af-46e4-9b44-83b9389a531b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.484288] env[62235]: DEBUG nova.scheduler.client.report [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.504285] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 554.504491] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 554.504669] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleting the datastore file [datastore1] d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 554.504914] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c668079-f933-4dc9-86fd-fdded16a8c07 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.512703] env[62235]: DEBUG oslo_vmware.api [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for the task: (returnval){ [ 554.512703] env[62235]: value = "task-1271454" [ 554.512703] env[62235]: _type = "Task" [ 554.512703] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.525540] env[62235]: DEBUG oslo_vmware.api [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.682312] env[62235]: DEBUG nova.network.neutron [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.993105] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.993665] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.997979] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.716s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.029546] env[62235]: DEBUG oslo_vmware.api [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Task: {'id': task-1271454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.105094} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.029799] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 555.030159] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 555.030159] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 555.030434] env[62235]: INFO nova.compute.manager [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Took 1.13 seconds to destroy the instance on the hypervisor. [ 555.030570] env[62235]: DEBUG oslo.service.loopingcall [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.030756] env[62235]: DEBUG nova.compute.manager [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.030847] env[62235]: DEBUG nova.network.neutron [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 555.055301] env[62235]: DEBUG nova.network.neutron [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.187755] env[62235]: INFO nova.compute.manager [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7] Took 1.06 seconds to deallocate network for instance. [ 555.499932] env[62235]: DEBUG nova.compute.utils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.505278] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 555.505278] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 555.559910] env[62235]: DEBUG nova.network.neutron [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.705531] env[62235]: DEBUG nova.policy [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dccfc771ad5a446982347200e7976828', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f1d954c9af3d4716893a8f1f3fafd20b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 556.008355] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.018553] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd0ed81-ee82-4ceb-a764-32c98d46461b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.030485] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019df954-0454-4aa2-b2d2-cf0826bea7f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.076725] env[62235]: INFO nova.compute.manager [-] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Took 1.05 seconds to deallocate network for instance. [ 556.080860] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce79ec9-0933-41bc-b252-80cd827b5bff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.089927] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Acquiring lock "1d1a57fc-6bc7-47a1-918e-8d23ac797d54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.090516] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Lock "1d1a57fc-6bc7-47a1-918e-8d23ac797d54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.096300] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b215b25b-6748-4f40-9596-ebbbdc1e76c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.117446] env[62235]: DEBUG nova.compute.provider_tree [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.239259] env[62235]: INFO nova.scheduler.client.report [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted allocations for instance 3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7 [ 556.460124] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Successfully created port: f1a29580-7286-4eab-bbcf-0149f95ed7b2 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.593931] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.621057] env[62235]: DEBUG nova.scheduler.client.report [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.750015] env[62235]: DEBUG oslo_concurrency.lockutils [None req-37b5217b-0860-458a-8793-7451250e66f4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "3d31c370-2dec-4aaf-bc98-8bf3cb6a52b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.048s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.038995] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.073111] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.073384] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.073503] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.073680] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.074176] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.074176] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.074578] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.074775] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.075233] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.075233] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.075371] env[62235]: DEBUG nova.virt.hardware [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.076269] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a66aae0-8c7d-41cc-b6f1-924782332bc0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.087949] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77673d5-a238-488a-bde6-0dd21cf3bf57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.125987] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.128s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.129102] env[62235]: ERROR nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Traceback (most recent call last): [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self.driver.spawn(context, instance, image_meta, [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] vm_ref = self.build_virtual_machine(instance, [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.129102] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] for vif in network_info: [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return self._sync_wrapper(fn, *args, **kwargs) [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self.wait() [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self[:] = self._gt.wait() [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return self._exit_event.wait() [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] result = hub.switch() [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.137031] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return self.greenlet.switch() [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] result = function(*args, **kwargs) [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] return func(*args, **kwargs) [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] raise e [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] nwinfo = self.network_api.allocate_for_instance( [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] created_port_ids = self._update_ports_for_instance( [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] with excutils.save_and_reraise_exception(): [ 557.137918] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] self.force_reraise() [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] raise self.value [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] updated_port = self._update_port( [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] _ensure_no_port_binding_failure(port) [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] raise exception.PortBindingFailed(port_id=port['id']) [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] nova.exception.PortBindingFailed: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. [ 557.138551] env[62235]: ERROR nova.compute.manager [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] [ 557.139337] env[62235]: DEBUG nova.compute.utils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.139337] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.761s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.139337] env[62235]: INFO nova.compute.claims [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 557.141249] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Build of instance d386f90d-25c3-4843-a91f-a092d27a6b5b was re-scheduled: Binding failed for port a66b3a5f-3385-48be-b820-39f6a3a54039, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.141658] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.141894] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.142515] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquired lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.142515] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.344036] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.705167] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.785030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.856993] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.362974] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Releasing lock "refresh_cache-d386f90d-25c3-4843-a91f-a092d27a6b5b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.363244] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 558.363453] env[62235]: DEBUG nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.363627] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.423813] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.497733] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Acquiring lock "60f868e1-e138-47d4-a510-5d3cd412e7de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.498211] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Lock "60f868e1-e138-47d4-a510-5d3cd412e7de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.600250] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7281ef71-782a-445c-bde7-9f06fdc0b2db {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.611676] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce3d187-ff16-4cd7-8064-296d3afd52e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.647930] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d943aeeb-325c-490f-b04d-2036a47cb108 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.656113] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93d7f04-b155-4195-93a8-8dba654c5719 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.676934] env[62235]: DEBUG nova.compute.provider_tree [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.929028] env[62235]: DEBUG nova.network.neutron [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.182950] env[62235]: DEBUG nova.scheduler.client.report [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.433660] env[62235]: INFO nova.compute.manager [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: d386f90d-25c3-4843-a91f-a092d27a6b5b] Took 1.07 seconds to deallocate network for instance. [ 559.687636] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.690431] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 559.692102] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.844s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.161733] env[62235]: DEBUG nova.compute.manager [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Received event network-changed-f1a29580-7286-4eab-bbcf-0149f95ed7b2 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.161970] env[62235]: DEBUG nova.compute.manager [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Refreshing instance network info cache due to event network-changed-f1a29580-7286-4eab-bbcf-0149f95ed7b2. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.162193] env[62235]: DEBUG oslo_concurrency.lockutils [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] Acquiring lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.162281] env[62235]: DEBUG oslo_concurrency.lockutils [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] Acquired lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.162714] env[62235]: DEBUG nova.network.neutron [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Refreshing network info cache for port f1a29580-7286-4eab-bbcf-0149f95ed7b2 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.198516] env[62235]: DEBUG nova.compute.utils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.207935] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 560.207935] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 560.217808] env[62235]: ERROR nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. [ 560.217808] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 560.217808] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.217808] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 560.217808] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.217808] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 560.217808] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.217808] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 560.217808] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.217808] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 560.217808] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.217808] env[62235]: ERROR nova.compute.manager raise self.value [ 560.217808] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.217808] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 560.217808] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.217808] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 560.218369] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.218369] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 560.218369] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. [ 560.218369] env[62235]: ERROR nova.compute.manager [ 560.218369] env[62235]: Traceback (most recent call last): [ 560.218369] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 560.218369] env[62235]: listener.cb(fileno) [ 560.218369] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.218369] env[62235]: result = function(*args, **kwargs) [ 560.218369] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.218369] env[62235]: return func(*args, **kwargs) [ 560.218369] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.218369] env[62235]: raise e [ 560.218369] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.218369] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 560.218369] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.218369] env[62235]: created_port_ids = self._update_ports_for_instance( [ 560.218369] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.218369] env[62235]: with excutils.save_and_reraise_exception(): [ 560.218369] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.218369] env[62235]: self.force_reraise() [ 560.218369] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.218369] env[62235]: raise self.value [ 560.218369] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.218369] env[62235]: updated_port = self._update_port( [ 560.218369] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.218369] env[62235]: _ensure_no_port_binding_failure(port) [ 560.218369] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.218369] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 560.219207] env[62235]: nova.exception.PortBindingFailed: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. [ 560.219207] env[62235]: Removing descriptor: 16 [ 560.219207] env[62235]: ERROR nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Traceback (most recent call last): [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] yield resources [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self.driver.spawn(context, instance, image_meta, [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.219207] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] vm_ref = self.build_virtual_machine(instance, [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] for vif in network_info: [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return self._sync_wrapper(fn, *args, **kwargs) [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self.wait() [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self[:] = self._gt.wait() [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return self._exit_event.wait() [ 560.219584] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] result = hub.switch() [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return self.greenlet.switch() [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] result = function(*args, **kwargs) [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return func(*args, **kwargs) [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] raise e [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] nwinfo = self.network_api.allocate_for_instance( [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.219986] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] created_port_ids = self._update_ports_for_instance( [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] with excutils.save_and_reraise_exception(): [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self.force_reraise() [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] raise self.value [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] updated_port = self._update_port( [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] _ensure_no_port_binding_failure(port) [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.220405] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] raise exception.PortBindingFailed(port_id=port['id']) [ 560.220762] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] nova.exception.PortBindingFailed: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. [ 560.220762] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] [ 560.220762] env[62235]: INFO nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Terminating instance [ 560.224367] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Acquiring lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.361727] env[62235]: DEBUG nova.policy [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02eba47eef1f456facd5e9859371e674', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a313826b6974e9992103979a52bba79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 560.473899] env[62235]: INFO nova.scheduler.client.report [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Deleted allocations for instance d386f90d-25c3-4843-a91f-a092d27a6b5b [ 560.690150] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be940c8-2e6a-4e00-a9ce-be56f6c1b39f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.698544] env[62235]: DEBUG nova.network.neutron [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.701242] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8cbccd-57aa-4b95-8d89-fe94e95c0e9f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.738244] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 560.743301] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0532e143-6ab1-472c-a311-daf6ad407b54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.754157] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793ba334-61e2-402d-ab77-83b2ea771973 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.770341] env[62235]: DEBUG nova.compute.provider_tree [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.936074] env[62235]: DEBUG nova.network.neutron [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.983975] env[62235]: DEBUG oslo_concurrency.lockutils [None req-986db6ac-d325-400d-a133-ee453a92e086 tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "d386f90d-25c3-4843-a91f-a092d27a6b5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.289s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.079931] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Successfully created port: 506492c3-fdb0-4d15-afef-31713c721d2b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 561.275287] env[62235]: DEBUG nova.scheduler.client.report [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.371762] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "8c4843d9-333f-4a08-9ec1-7971ecf53875" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.371987] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "8c4843d9-333f-4a08-9ec1-7971ecf53875" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.439615] env[62235]: DEBUG oslo_concurrency.lockutils [req-8a26c88c-8aaf-425f-8e06-bf6506c81ae8 req-773b639a-b399-4db4-82bc-88924692c687 service nova] Releasing lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.439615] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Acquired lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.439615] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 561.488499] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.655508] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Acquiring lock "f830c92d-ccf3-4612-980f-24ed8f679d4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.655749] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Lock "f830c92d-ccf3-4612-980f-24ed8f679d4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.749557] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 561.781914] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.782422] env[62235]: ERROR nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Traceback (most recent call last): [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self.driver.spawn(context, instance, image_meta, [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] vm_ref = self.build_virtual_machine(instance, [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.782422] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] for vif in network_info: [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return self._sync_wrapper(fn, *args, **kwargs) [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self.wait() [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self[:] = self._gt.wait() [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return self._exit_event.wait() [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] result = hub.switch() [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.783231] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return self.greenlet.switch() [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] result = function(*args, **kwargs) [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] return func(*args, **kwargs) [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] raise e [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] nwinfo = self.network_api.allocate_for_instance( [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] created_port_ids = self._update_ports_for_instance( [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] with excutils.save_and_reraise_exception(): [ 561.783929] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] self.force_reraise() [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] raise self.value [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] updated_port = self._update_port( [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] _ensure_no_port_binding_failure(port) [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] raise exception.PortBindingFailed(port_id=port['id']) [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] nova.exception.PortBindingFailed: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. [ 561.784315] env[62235]: ERROR nova.compute.manager [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] [ 561.785453] env[62235]: DEBUG nova.compute.utils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 561.789628] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 561.789628] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 561.789628] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 561.790289] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 561.790289] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 561.790289] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 561.790289] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 561.790289] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 561.790514] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 561.790514] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 561.790514] env[62235]: DEBUG nova.virt.hardware [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 561.790514] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.672s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.794196] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee593e59-6adb-450e-91ee-da9c37295b54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.797774] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Build of instance db07ab9b-13a6-42e5-bc60-1a1919696f0e was re-scheduled: Binding failed for port 7ba7ac15-7084-4fbe-ae35-80069dc46a3f, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 561.798998] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 561.798998] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Acquiring lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.798998] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Acquired lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.798998] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 561.807460] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8da584-a4ce-490d-9539-21790854e9da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.964163] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.026029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.182972] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.330895] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.652312] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Acquiring lock "584b41c1-976d-49e9-b5ea-c5c22e1877c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.652312] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Lock "584b41c1-976d-49e9-b5ea-c5c22e1877c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.654628] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac421c6-3df1-4b53-bc23-7b05ba4dde1d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.663305] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3260460-5f75-470d-b5a6-bf5f1f49b470 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.695456] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Releasing lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.696075] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 562.696478] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 562.696996] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54a2e780-8c02-41f1-afcd-786c8a209e04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.699445] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d010a7a-f1b8-4352-b554-c6dd73f4c091 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.708065] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.710431] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4bc774-0fc7-4c1f-90e6-1d3641bb9050 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.717916] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fade44d0-eaab-4c0f-993a-14a7640d6361 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.743067] env[62235]: DEBUG nova.compute.provider_tree [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.748359] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf could not be found. [ 562.748359] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 562.748497] env[62235]: INFO nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 562.748715] env[62235]: DEBUG oslo.service.loopingcall [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.749161] env[62235]: DEBUG nova.compute.manager [-] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.749161] env[62235]: DEBUG nova.network.neutron [-] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.774480] env[62235]: DEBUG nova.network.neutron [-] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.183629] env[62235]: DEBUG nova.compute.manager [req-d1226699-18ca-4afb-92b0-9ec0967d0458 req-10934568-1c74-48c5-84ec-2e9eb0d4e523 service nova] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Received event network-vif-deleted-f1a29580-7286-4eab-bbcf-0149f95ed7b2 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.216891] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Releasing lock "refresh_cache-db07ab9b-13a6-42e5-bc60-1a1919696f0e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.216891] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 563.216891] env[62235]: DEBUG nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.216891] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.250547] env[62235]: DEBUG nova.scheduler.client.report [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.258298] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.279074] env[62235]: DEBUG nova.network.neutron [-] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.289842] env[62235]: ERROR nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. [ 563.289842] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.289842] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.289842] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.289842] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.289842] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.289842] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.289842] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.289842] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.289842] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 563.289842] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.289842] env[62235]: ERROR nova.compute.manager raise self.value [ 563.289842] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.289842] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.289842] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.289842] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.290344] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.290344] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.290344] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. [ 563.290344] env[62235]: ERROR nova.compute.manager [ 563.290344] env[62235]: Traceback (most recent call last): [ 563.290344] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.290344] env[62235]: listener.cb(fileno) [ 563.290344] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.290344] env[62235]: result = function(*args, **kwargs) [ 563.290344] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.290344] env[62235]: return func(*args, **kwargs) [ 563.290344] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.290344] env[62235]: raise e [ 563.290344] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.290344] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 563.290344] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.290344] env[62235]: created_port_ids = self._update_ports_for_instance( [ 563.290344] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.290344] env[62235]: with excutils.save_and_reraise_exception(): [ 563.290344] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.290344] env[62235]: self.force_reraise() [ 563.290344] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.290344] env[62235]: raise self.value [ 563.290344] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.290344] env[62235]: updated_port = self._update_port( [ 563.290344] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.290344] env[62235]: _ensure_no_port_binding_failure(port) [ 563.290344] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.290344] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.291335] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. [ 563.291335] env[62235]: Removing descriptor: 22 [ 563.291335] env[62235]: ERROR nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] Traceback (most recent call last): [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] yield resources [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self.driver.spawn(context, instance, image_meta, [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.291335] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] vm_ref = self.build_virtual_machine(instance, [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] for vif in network_info: [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return self._sync_wrapper(fn, *args, **kwargs) [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self.wait() [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self[:] = self._gt.wait() [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return self._exit_event.wait() [ 563.291778] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] result = hub.switch() [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return self.greenlet.switch() [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] result = function(*args, **kwargs) [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return func(*args, **kwargs) [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] raise e [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] nwinfo = self.network_api.allocate_for_instance( [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.292220] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] created_port_ids = self._update_ports_for_instance( [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] with excutils.save_and_reraise_exception(): [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self.force_reraise() [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] raise self.value [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] updated_port = self._update_port( [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] _ensure_no_port_binding_failure(port) [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.292618] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] raise exception.PortBindingFailed(port_id=port['id']) [ 563.292968] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] nova.exception.PortBindingFailed: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. [ 563.292968] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] [ 563.292968] env[62235]: INFO nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Terminating instance [ 563.295623] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Acquiring lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.295794] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Acquired lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.295966] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.593809] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "4d10be8a-a4ba-46f7-b4da-c37f8520bab6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.594576] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "4d10be8a-a4ba-46f7-b4da-c37f8520bab6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.637737] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Acquiring lock "2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.637961] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Lock "2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.763021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.763021] env[62235]: ERROR nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. [ 563.763021] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Traceback (most recent call last): [ 563.763021] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.763021] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self.driver.spawn(context, instance, image_meta, [ 563.763021] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 563.763021] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.763021] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.763021] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] vm_ref = self.build_virtual_machine(instance, [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] for vif in network_info: [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return self._sync_wrapper(fn, *args, **kwargs) [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self.wait() [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self[:] = self._gt.wait() [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return self._exit_event.wait() [ 563.763489] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] result = hub.switch() [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return self.greenlet.switch() [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] result = function(*args, **kwargs) [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] return func(*args, **kwargs) [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] raise e [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] nwinfo = self.network_api.allocate_for_instance( [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.764023] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] created_port_ids = self._update_ports_for_instance( [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] with excutils.save_and_reraise_exception(): [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] self.force_reraise() [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] raise self.value [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] updated_port = self._update_port( [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] _ensure_no_port_binding_failure(port) [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.764538] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] raise exception.PortBindingFailed(port_id=port['id']) [ 563.765052] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] nova.exception.PortBindingFailed: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. [ 563.765052] env[62235]: ERROR nova.compute.manager [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] [ 563.765052] env[62235]: DEBUG nova.compute.utils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 563.769063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.646s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.769063] env[62235]: DEBUG nova.objects.instance [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lazy-loading 'resources' on Instance uuid 4e20ae82-9a11-433b-ac4e-a85404cf4ec6 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 563.769063] env[62235]: DEBUG nova.network.neutron [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.769469] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Build of instance dbba80e1-52e2-4787-bcf7-c3300e923813 was re-scheduled: Binding failed for port 6acfaaad-2083-412d-8415-8cd96596e71d, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 563.770148] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 563.770539] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Acquiring lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.770837] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Acquired lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.772100] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.781305] env[62235]: INFO nova.compute.manager [-] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Took 1.03 seconds to deallocate network for instance. [ 563.784019] env[62235]: DEBUG nova.compute.claims [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 563.784019] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.819530] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.941375] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.283025] env[62235]: INFO nova.compute.manager [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] [instance: db07ab9b-13a6-42e5-bc60-1a1919696f0e] Took 1.06 seconds to deallocate network for instance. [ 564.332279] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.447659] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Releasing lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.448093] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 564.448290] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.448807] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3893dc86-4278-48db-9f94-a0a77107e228 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.461621] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45b0993-d792-4d48-a6e2-b767ce96c037 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.476488] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.490881] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 842f0697-6cec-423b-a35a-583985713a08 could not be found. [ 564.491112] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.491311] env[62235]: INFO nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Took 0.04 seconds to destroy the instance on the hypervisor. [ 564.491558] env[62235]: DEBUG oslo.service.loopingcall [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.492404] env[62235]: DEBUG nova.compute.manager [-] [instance: 842f0697-6cec-423b-a35a-583985713a08] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.492526] env[62235]: DEBUG nova.network.neutron [-] [instance: 842f0697-6cec-423b-a35a-583985713a08] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.521256] env[62235]: DEBUG nova.network.neutron [-] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.774980] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b471b5-f9fe-4686-ac2c-abce436df968 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.785538] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1b6c06-f9a3-46d5-92a0-6b47df360516 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.827021] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0356816-480d-4201-9265-4b70f99f8a5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.836285] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6ff88a-c77c-413f-bb76-872307975d58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.855592] env[62235]: DEBUG nova.compute.provider_tree [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.890113] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "3f245678-41f4-4685-87fc-822bf877d43e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.890362] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "3f245678-41f4-4685-87fc-822bf877d43e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.988467] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Releasing lock "refresh_cache-dbba80e1-52e2-4787-bcf7-c3300e923813" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.988467] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 564.988467] env[62235]: DEBUG nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.988467] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.004658] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.024346] env[62235]: DEBUG nova.network.neutron [-] [instance: 842f0697-6cec-423b-a35a-583985713a08] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.230080] env[62235]: DEBUG nova.compute.manager [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] [instance: 842f0697-6cec-423b-a35a-583985713a08] Received event network-changed-506492c3-fdb0-4d15-afef-31713c721d2b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.230080] env[62235]: DEBUG nova.compute.manager [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] [instance: 842f0697-6cec-423b-a35a-583985713a08] Refreshing instance network info cache due to event network-changed-506492c3-fdb0-4d15-afef-31713c721d2b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.230080] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] Acquiring lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.230080] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] Acquired lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.232197] env[62235]: DEBUG nova.network.neutron [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] [instance: 842f0697-6cec-423b-a35a-583985713a08] Refreshing network info cache for port 506492c3-fdb0-4d15-afef-31713c721d2b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 565.333303] env[62235]: INFO nova.scheduler.client.report [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Deleted allocations for instance db07ab9b-13a6-42e5-bc60-1a1919696f0e [ 565.360450] env[62235]: DEBUG nova.scheduler.client.report [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.509375] env[62235]: DEBUG nova.network.neutron [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.531124] env[62235]: INFO nova.compute.manager [-] [instance: 842f0697-6cec-423b-a35a-583985713a08] Took 1.04 seconds to deallocate network for instance. [ 565.534831] env[62235]: DEBUG nova.compute.claims [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.535400] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.771180] env[62235]: DEBUG nova.network.neutron [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.846119] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6db9c68a-d242-4a4d-a585-ca244f06f4d5 tempest-AttachInterfacesUnderV243Test-915359420 tempest-AttachInterfacesUnderV243Test-915359420-project-member] Lock "db07ab9b-13a6-42e5-bc60-1a1919696f0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.638s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.867655] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.102s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.869763] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.420s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.902494] env[62235]: INFO nova.scheduler.client.report [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Deleted allocations for instance 4e20ae82-9a11-433b-ac4e-a85404cf4ec6 [ 565.923175] env[62235]: DEBUG nova.network.neutron [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] [instance: 842f0697-6cec-423b-a35a-583985713a08] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.015737] env[62235]: INFO nova.compute.manager [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] [instance: dbba80e1-52e2-4787-bcf7-c3300e923813] Took 1.03 seconds to deallocate network for instance. [ 566.348818] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 566.411587] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9b28bb76-722d-4217-a169-2be1e63d6ec3 tempest-ServerDiagnosticsV248Test-1352630323 tempest-ServerDiagnosticsV248Test-1352630323-project-member] Lock "4e20ae82-9a11-433b-ac4e-a85404cf4ec6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.139s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.427905] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] Releasing lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.428111] env[62235]: DEBUG nova.compute.manager [req-1b966df5-472a-4c44-85be-e0929213008c req-fed8643d-6deb-4e05-9040-cc393a66b89a service nova] [instance: 842f0697-6cec-423b-a35a-583985713a08] Received event network-vif-deleted-506492c3-fdb0-4d15-afef-31713c721d2b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.889140] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.909657] env[62235]: WARNING nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 567.049290] env[62235]: INFO nova.scheduler.client.report [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Deleted allocations for instance dbba80e1-52e2-4787-bcf7-c3300e923813 [ 567.412395] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance dbba80e1-52e2-4787-bcf7-c3300e923813 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.412549] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 7d0facf3-aa88-4332-8077-418db20be8b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 567.412672] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 567.412790] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 842f0697-6cec-423b-a35a-583985713a08 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 567.543962] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Acquiring lock "19e40767-d9f5-4b78-80f0-7f0f24464d40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.544222] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Lock "19e40767-d9f5-4b78-80f0-7f0f24464d40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.559522] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a96cc978-410d-42aa-9dfc-5451be94ef63 tempest-ServersTestJSON-1744534731 tempest-ServersTestJSON-1744534731-project-member] Lock "dbba80e1-52e2-4787-bcf7-c3300e923813" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.990s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.917352] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ab123e55-fa0e-4216-9afc-f8672598b08e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.068143] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 568.421604] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b6122fc8-989d-4321-ac7a-80b2402a695f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.598137] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.925174] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance c8efd1fa-0e33-423c-975e-b2583e1cfbd4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.427972] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 028e94f4-6b8f-45c3-a046-ed44e575362a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.932689] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 41adaeb9-c39f-4214-b9d1-7137aab3455f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 570.436481] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance e7684530-e42d-4a30-9c7c-28be8ff31b7b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 570.941493] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ec00f2ba-ed0c-47de-a9b2-bd277f96e65f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 571.444793] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a84e793b-4dbd-4b89-8fc4-94baacecd8e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 571.951399] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 063af6b5-92ce-40b0-81e0-0dd6155898fc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 572.342320] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "b348bc9f-3bd4-475f-9957-42a99848005c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.342566] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "b348bc9f-3bd4-475f-9957-42a99848005c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.455438] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 7adac743-926e-4c66-8fbe-02b96fbdb2b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 572.960143] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 74d069ed-c262-4b91-bcdd-ea3b4121adc4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 572.999827] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "012dd63b-a092-4a1d-a79c-1e1bf2fd8771" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.999827] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "012dd63b-a092-4a1d-a79c-1e1bf2fd8771" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.463710] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 573.967747] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 91e05f87-96a4-4d52-9016-841a3a034930 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 574.473072] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 279097d2-348d-4f58-88ec-8e2e3143f754 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 574.863160] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "be4fb088-8b01-4245-82fb-b0e085825597" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.863707] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "be4fb088-8b01-4245-82fb-b0e085825597" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.974765] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 1d1a57fc-6bc7-47a1-918e-8d23ac797d54 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.479169] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 60f868e1-e138-47d4-a510-5d3cd412e7de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.990081] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 8c4843d9-333f-4a08-9ec1-7971ecf53875 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.493471] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance f830c92d-ccf3-4612-980f-24ed8f679d4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.998108] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 584b41c1-976d-49e9-b5ea-c5c22e1877c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.501109] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 4d10be8a-a4ba-46f7-b4da-c37f8520bab6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.012334] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.015906] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Acquiring lock "693db195-ce52-40ff-bbb4-91f03d5a0bd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.015906] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Lock "693db195-ce52-40ff-bbb4-91f03d5a0bd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.457513] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "452cf04e-d6e9-48a8-a77b-f931c9440858" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.457513] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "452cf04e-d6e9-48a8-a77b-f931c9440858" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.517690] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3f245678-41f4-4685-87fc-822bf877d43e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.517690] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 578.517690] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 579.007816] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963fe3e3-fea1-4ea7-a6b1-0f9ff914dd2b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.015609] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea975660-23c6-4636-8ba3-c1ba6f12370a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.045486] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163be4fc-937b-42eb-a1d3-758d5ea5fca0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.053529] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f90f3e4-9213-4756-b2ff-cd3f1447c3be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.068171] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.571135] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.076395] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 580.076765] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.207s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.076915] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.925s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.030083] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302de277-401a-486b-9775-e8d92cad5a5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.038713] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d40fa3-c03a-4314-a30b-a8b239a4e9f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.071439] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d46b138-4b92-4bf7-b96e-f0e979c9060e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.079947] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc00aa3-87ed-4459-aca9-3997104a6a81 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.094315] env[62235]: DEBUG nova.compute.provider_tree [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.597410] env[62235]: DEBUG nova.scheduler.client.report [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.102958] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.026s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.103635] env[62235]: ERROR nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Traceback (most recent call last): [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self.driver.spawn(context, instance, image_meta, [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] vm_ref = self.build_virtual_machine(instance, [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.103635] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] for vif in network_info: [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return self._sync_wrapper(fn, *args, **kwargs) [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self.wait() [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self[:] = self._gt.wait() [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return self._exit_event.wait() [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] result = hub.switch() [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.103994] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return self.greenlet.switch() [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] result = function(*args, **kwargs) [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] return func(*args, **kwargs) [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] raise e [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] nwinfo = self.network_api.allocate_for_instance( [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] created_port_ids = self._update_ports_for_instance( [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] with excutils.save_and_reraise_exception(): [ 582.104432] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] self.force_reraise() [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] raise self.value [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] updated_port = self._update_port( [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] _ensure_no_port_binding_failure(port) [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] raise exception.PortBindingFailed(port_id=port['id']) [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] nova.exception.PortBindingFailed: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. [ 582.104865] env[62235]: ERROR nova.compute.manager [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] [ 582.105322] env[62235]: DEBUG nova.compute.utils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.106965] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.775s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.107124] env[62235]: DEBUG nova.objects.instance [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 582.109651] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Build of instance 7d0facf3-aa88-4332-8077-418db20be8b4 was re-scheduled: Binding failed for port 4b47ecd1-f586-4dc0-9b2b-bc6926115db9, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.110090] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.110314] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.110455] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquired lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.110607] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.632740] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.720422] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.120091] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e3d68dab-528d-4872-970a-d68f3c5c8d88 tempest-ServersAdmin275Test-80944544 tempest-ServersAdmin275Test-80944544-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.121243] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.368s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.125092] env[62235]: INFO nova.compute.claims [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.226104] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Releasing lock "refresh_cache-7d0facf3-aa88-4332-8077-418db20be8b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.226104] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.226104] env[62235]: DEBUG nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.226430] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.242090] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.745105] env[62235]: DEBUG nova.network.neutron [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.247619] env[62235]: INFO nova.compute.manager [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 7d0facf3-aa88-4332-8077-418db20be8b4] Took 1.02 seconds to deallocate network for instance. [ 584.496236] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9851586a-4653-4409-a96f-0088bd2a1548 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.503817] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d668374d-0087-435e-829b-1b6b414fd480 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.533549] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692c296e-36fc-4125-8f08-30ecb13fe5ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.540869] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc264a13-6911-431f-aca2-fc1495f70d49 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.553920] env[62235]: DEBUG nova.compute.provider_tree [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.056361] env[62235]: DEBUG nova.scheduler.client.report [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.298938] env[62235]: INFO nova.scheduler.client.report [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Deleted allocations for instance 7d0facf3-aa88-4332-8077-418db20be8b4 [ 585.562787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.562787] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.564435] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.620s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.565963] env[62235]: INFO nova.compute.claims [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.811231] env[62235]: DEBUG oslo_concurrency.lockutils [None req-40904f2e-f962-4b6f-9576-062e2ab7894e tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "7d0facf3-aa88-4332-8077-418db20be8b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.774s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.072820] env[62235]: DEBUG nova.compute.utils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.074238] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 586.075037] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 586.140860] env[62235]: DEBUG nova.policy [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '208e97254d194d37ac2aa971d7a090c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b8663bd2607421bb22c719ea94cd989', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 586.314017] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.570544] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Successfully created port: 064a1463-1523-4a54-9bd7-c0e5d15800e0 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.577601] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.716112] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "1c6f4360-b80b-43e3-accc-f7500e626100" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.716742] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "1c6f4360-b80b-43e3-accc-f7500e626100" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.839494] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.032335] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b796111-f977-4d0b-b778-d42a13da290f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.040839] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03431339-c1e4-4f32-9f49-d37379cae3ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.073052] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d067d825-d566-48dc-aaaf-c021f35646ff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.080498] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171048b1-c6bf-49d6-b9df-b5cdad4d3e76 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.097946] env[62235]: DEBUG nova.compute.provider_tree [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.319172] env[62235]: DEBUG nova.compute.manager [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Received event network-changed-064a1463-1523-4a54-9bd7-c0e5d15800e0 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.319404] env[62235]: DEBUG nova.compute.manager [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Refreshing instance network info cache due to event network-changed-064a1463-1523-4a54-9bd7-c0e5d15800e0. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 587.319515] env[62235]: DEBUG oslo_concurrency.lockutils [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] Acquiring lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.319635] env[62235]: DEBUG oslo_concurrency.lockutils [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] Acquired lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.319803] env[62235]: DEBUG nova.network.neutron [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Refreshing network info cache for port 064a1463-1523-4a54-9bd7-c0e5d15800e0 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.533959] env[62235]: ERROR nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. [ 587.533959] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.533959] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.533959] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.533959] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.533959] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.533959] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.533959] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.533959] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.533959] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 587.533959] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.533959] env[62235]: ERROR nova.compute.manager raise self.value [ 587.533959] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.533959] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.533959] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.533959] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.534457] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.534457] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.534457] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. [ 587.534457] env[62235]: ERROR nova.compute.manager [ 587.534457] env[62235]: Traceback (most recent call last): [ 587.534457] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.534457] env[62235]: listener.cb(fileno) [ 587.534457] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.534457] env[62235]: result = function(*args, **kwargs) [ 587.534457] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.534457] env[62235]: return func(*args, **kwargs) [ 587.534457] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.534457] env[62235]: raise e [ 587.534457] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.534457] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 587.534457] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.534457] env[62235]: created_port_ids = self._update_ports_for_instance( [ 587.534457] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.534457] env[62235]: with excutils.save_and_reraise_exception(): [ 587.534457] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.534457] env[62235]: self.force_reraise() [ 587.534457] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.534457] env[62235]: raise self.value [ 587.534457] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.534457] env[62235]: updated_port = self._update_port( [ 587.534457] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.534457] env[62235]: _ensure_no_port_binding_failure(port) [ 587.534457] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.534457] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.535352] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. [ 587.535352] env[62235]: Removing descriptor: 22 [ 587.589766] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.601038] env[62235]: DEBUG nova.scheduler.client.report [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.622316] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.622552] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.622703] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.622881] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.623032] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.623178] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.623377] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.623530] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.623690] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.623844] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.624017] env[62235]: DEBUG nova.virt.hardware [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.625148] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8f78d6-da72-49e2-a6eb-0bc1bc1b8908 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.633982] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b25a95-e287-4dcc-9aa8-cd49b9d5943e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.651083] env[62235]: ERROR nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Traceback (most recent call last): [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] yield resources [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self.driver.spawn(context, instance, image_meta, [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] vm_ref = self.build_virtual_machine(instance, [ 587.651083] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] for vif in network_info: [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] return self._sync_wrapper(fn, *args, **kwargs) [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self.wait() [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self[:] = self._gt.wait() [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] return self._exit_event.wait() [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 587.651663] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] current.throw(*self._exc) [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] result = function(*args, **kwargs) [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] return func(*args, **kwargs) [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] raise e [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] nwinfo = self.network_api.allocate_for_instance( [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] created_port_ids = self._update_ports_for_instance( [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] with excutils.save_and_reraise_exception(): [ 587.652257] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self.force_reraise() [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] raise self.value [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] updated_port = self._update_port( [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] _ensure_no_port_binding_failure(port) [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] raise exception.PortBindingFailed(port_id=port['id']) [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] nova.exception.PortBindingFailed: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. [ 587.652662] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] [ 587.652662] env[62235]: INFO nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Terminating instance [ 587.654200] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.838571] env[62235]: DEBUG nova.network.neutron [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.901988] env[62235]: DEBUG nova.network.neutron [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.106931] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.107520] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.110411] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.638s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.111946] env[62235]: INFO nova.compute.claims [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.404136] env[62235]: DEBUG oslo_concurrency.lockutils [req-17ea02a4-6702-46de-b550-7f64a53b3054 req-b0956a77-2547-4279-b846-f6f2caf62d49 service nova] Releasing lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.404583] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquired lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.404777] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.617879] env[62235]: DEBUG nova.compute.utils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.620295] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.620295] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.657192] env[62235]: DEBUG nova.policy [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8dab53ec60dc40b1a6ff0c64ef85e64e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fef119795b40491b9468d88dc41ccfe4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.933986] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.024544] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Successfully created port: 922b5a90-358d-4fee-8bcf-5087b4e9e1be {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.028008] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.121444] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.359553] env[62235]: DEBUG nova.compute.manager [req-fdbdca35-ff34-4e8d-afa4-aa269fe94883 req-a1e74dcb-841c-4546-bea1-54871a86c536 service nova] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Received event network-vif-deleted-064a1463-1523-4a54-9bd7-c0e5d15800e0 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.531807] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Releasing lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.532302] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.532545] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.532895] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ce02210-8ac2-41ec-aa28-bd727da05f04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.553582] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ede0903-fd63-4a9b-afde-4fc3971c97e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.597635] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab123e55-fa0e-4216-9afc-f8672598b08e could not be found. [ 589.597635] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.597635] env[62235]: INFO nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Took 0.07 seconds to destroy the instance on the hypervisor. [ 589.599475] env[62235]: DEBUG oslo.service.loopingcall [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.602234] env[62235]: DEBUG nova.compute.manager [-] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.602234] env[62235]: DEBUG nova.network.neutron [-] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.636508] env[62235]: DEBUG nova.network.neutron [-] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.649368] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9190b451-76cb-40db-913e-292414c95019 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.659692] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60abfdb-e9b0-43fb-8118-00574ae64454 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.691992] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b68853-9e2d-4754-8e44-889738cd58d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.699936] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e1bf76-148e-4835-8342-cd36b872d9f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.713541] env[62235]: DEBUG nova.compute.provider_tree [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.115040] env[62235]: ERROR nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. [ 590.115040] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 590.115040] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.115040] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 590.115040] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.115040] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 590.115040] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.115040] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 590.115040] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.115040] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 590.115040] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.115040] env[62235]: ERROR nova.compute.manager raise self.value [ 590.115040] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.115040] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 590.115040] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.115040] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 590.115549] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.115549] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 590.115549] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. [ 590.115549] env[62235]: ERROR nova.compute.manager [ 590.115549] env[62235]: Traceback (most recent call last): [ 590.115549] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 590.115549] env[62235]: listener.cb(fileno) [ 590.115549] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.115549] env[62235]: result = function(*args, **kwargs) [ 590.115549] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.115549] env[62235]: return func(*args, **kwargs) [ 590.115549] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.115549] env[62235]: raise e [ 590.115549] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.115549] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 590.115549] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.115549] env[62235]: created_port_ids = self._update_ports_for_instance( [ 590.115549] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.115549] env[62235]: with excutils.save_and_reraise_exception(): [ 590.115549] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.115549] env[62235]: self.force_reraise() [ 590.115549] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.115549] env[62235]: raise self.value [ 590.115549] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.115549] env[62235]: updated_port = self._update_port( [ 590.115549] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.115549] env[62235]: _ensure_no_port_binding_failure(port) [ 590.115549] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.115549] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 590.116444] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. [ 590.116444] env[62235]: Removing descriptor: 22 [ 590.132373] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.140975] env[62235]: DEBUG nova.network.neutron [-] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.156945] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:18:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='9741002',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1788794746',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.157103] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.157233] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.157416] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.157558] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.157714] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.157901] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.158073] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.158243] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.158403] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.158570] env[62235]: DEBUG nova.virt.hardware [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.159418] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ff86c0-d986-4691-bae2-c70301401433 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.167783] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cd6a68-de56-4670-95cb-e4c3de091341 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.183607] env[62235]: ERROR nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Traceback (most recent call last): [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] yield resources [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self.driver.spawn(context, instance, image_meta, [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] vm_ref = self.build_virtual_machine(instance, [ 590.183607] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] for vif in network_info: [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] return self._sync_wrapper(fn, *args, **kwargs) [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self.wait() [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self[:] = self._gt.wait() [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] return self._exit_event.wait() [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 590.184114] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] current.throw(*self._exc) [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] result = function(*args, **kwargs) [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] return func(*args, **kwargs) [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] raise e [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] nwinfo = self.network_api.allocate_for_instance( [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] created_port_ids = self._update_ports_for_instance( [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] with excutils.save_and_reraise_exception(): [ 590.184565] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self.force_reraise() [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] raise self.value [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] updated_port = self._update_port( [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] _ensure_no_port_binding_failure(port) [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] raise exception.PortBindingFailed(port_id=port['id']) [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] nova.exception.PortBindingFailed: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. [ 590.185040] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] [ 590.185040] env[62235]: INFO nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Terminating instance [ 590.188730] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Acquiring lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.188893] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Acquired lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.189829] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 590.215933] env[62235]: DEBUG nova.scheduler.client.report [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.643687] env[62235]: INFO nova.compute.manager [-] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Took 1.04 seconds to deallocate network for instance. [ 590.646094] env[62235]: DEBUG nova.compute.claims [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.646274] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.706355] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.721972] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.722513] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.725031] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.132s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.725222] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.727150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.943s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.728513] env[62235]: INFO nova.compute.claims [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.778730] env[62235]: INFO nova.scheduler.client.report [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Deleted allocations for instance d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0 [ 590.820311] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.233985] env[62235]: DEBUG nova.compute.utils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.237820] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.237820] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.286694] env[62235]: DEBUG oslo_concurrency.lockutils [None req-043b9e16-2cc0-42e5-b81a-d80a73b0457c tempest-ServersAdmin275Test-1741558420 tempest-ServersAdmin275Test-1741558420-project-member] Lock "d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.013s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.297829] env[62235]: DEBUG nova.policy [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '208e97254d194d37ac2aa971d7a090c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b8663bd2607421bb22c719ea94cd989', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.322851] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Releasing lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.323343] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.323535] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 591.323819] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18fd9e07-4f42-4e04-9322-2db0b07aa95b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.333035] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1483f66-4bda-444b-86c1-05f69b35adaf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.354179] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6122fc8-989d-4321-ac7a-80b2402a695f could not be found. [ 591.354401] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 591.354583] env[62235]: INFO nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 591.354819] env[62235]: DEBUG oslo.service.loopingcall [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.355045] env[62235]: DEBUG nova.compute.manager [-] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.355142] env[62235]: DEBUG nova.network.neutron [-] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.373841] env[62235]: DEBUG nova.network.neutron [-] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.384158] env[62235]: DEBUG nova.compute.manager [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Received event network-changed-922b5a90-358d-4fee-8bcf-5087b4e9e1be {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.384342] env[62235]: DEBUG nova.compute.manager [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Refreshing instance network info cache due to event network-changed-922b5a90-358d-4fee-8bcf-5087b4e9e1be. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.384543] env[62235]: DEBUG oslo_concurrency.lockutils [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] Acquiring lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.384678] env[62235]: DEBUG oslo_concurrency.lockutils [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] Acquired lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.384828] env[62235]: DEBUG nova.network.neutron [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Refreshing network info cache for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 591.587787] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Successfully created port: 5a8cdc37-d4e8-4790-b051-82270896cc2c {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.738805] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.876443] env[62235]: DEBUG nova.network.neutron [-] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.909118] env[62235]: DEBUG nova.network.neutron [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.055364] env[62235]: DEBUG nova.network.neutron [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.169016] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a238b01-8b91-4c0a-b0e7-d267babd8a0d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.176150] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fc7f5e-ae78-416b-b537-41ed8c84a734 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.210809] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aee0a18-66e5-4405-9920-8dd6bd7fe326 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.221639] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a04e330-3bf2-452e-bc05-0cf7bf7f7b27 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.236191] env[62235]: DEBUG nova.compute.provider_tree [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.379549] env[62235]: INFO nova.compute.manager [-] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Took 1.02 seconds to deallocate network for instance. [ 592.382252] env[62235]: DEBUG nova.compute.claims [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.382561] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.558638] env[62235]: DEBUG oslo_concurrency.lockutils [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] Releasing lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.558638] env[62235]: DEBUG nova.compute.manager [req-da351d84-1459-4e82-92cd-1deefc5793fc req-e4b09a77-df10-4973-9f17-eb98266254ea service nova] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Received event network-vif-deleted-922b5a90-358d-4fee-8bcf-5087b4e9e1be {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.740223] env[62235]: DEBUG nova.scheduler.client.report [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.752231] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.784339] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.784657] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.784811] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.785058] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.785256] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.785427] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.785678] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.785842] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.786064] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.786269] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.786449] env[62235]: DEBUG nova.virt.hardware [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.790114] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdaa3a2e-7245-4cba-96e6-22acc1b93b51 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.798375] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7474b2-cf80-4ce8-8398-ee1c85ecc974 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.817948] env[62235]: ERROR nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. [ 592.817948] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.817948] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.817948] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.817948] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.817948] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.817948] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.817948] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.817948] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.817948] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 592.817948] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.817948] env[62235]: ERROR nova.compute.manager raise self.value [ 592.817948] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.817948] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.817948] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.817948] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.818887] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.818887] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.818887] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. [ 592.818887] env[62235]: ERROR nova.compute.manager [ 592.818887] env[62235]: Traceback (most recent call last): [ 592.818887] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.818887] env[62235]: listener.cb(fileno) [ 592.818887] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.818887] env[62235]: result = function(*args, **kwargs) [ 592.818887] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.818887] env[62235]: return func(*args, **kwargs) [ 592.818887] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.818887] env[62235]: raise e [ 592.818887] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.818887] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 592.818887] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.818887] env[62235]: created_port_ids = self._update_ports_for_instance( [ 592.818887] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.818887] env[62235]: with excutils.save_and_reraise_exception(): [ 592.818887] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.818887] env[62235]: self.force_reraise() [ 592.818887] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.818887] env[62235]: raise self.value [ 592.818887] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.818887] env[62235]: updated_port = self._update_port( [ 592.818887] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.818887] env[62235]: _ensure_no_port_binding_failure(port) [ 592.818887] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.818887] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.821253] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. [ 592.821253] env[62235]: Removing descriptor: 22 [ 592.821253] env[62235]: ERROR nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Traceback (most recent call last): [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] yield resources [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self.driver.spawn(context, instance, image_meta, [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.821253] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] vm_ref = self.build_virtual_machine(instance, [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] for vif in network_info: [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return self._sync_wrapper(fn, *args, **kwargs) [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self.wait() [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self[:] = self._gt.wait() [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return self._exit_event.wait() [ 592.821696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] result = hub.switch() [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return self.greenlet.switch() [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] result = function(*args, **kwargs) [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return func(*args, **kwargs) [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] raise e [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] nwinfo = self.network_api.allocate_for_instance( [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.822223] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] created_port_ids = self._update_ports_for_instance( [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] with excutils.save_and_reraise_exception(): [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self.force_reraise() [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] raise self.value [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] updated_port = self._update_port( [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] _ensure_no_port_binding_failure(port) [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.822655] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] raise exception.PortBindingFailed(port_id=port['id']) [ 592.823337] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] nova.exception.PortBindingFailed: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. [ 592.823337] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] [ 592.823337] env[62235]: INFO nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Terminating instance [ 592.823337] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.823337] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquired lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.823337] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.245008] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.245551] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.248138] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.223s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.252168] env[62235]: INFO nova.compute.claims [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.344694] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.407932] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.416634] env[62235]: DEBUG nova.compute.manager [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Received event network-changed-5a8cdc37-d4e8-4790-b051-82270896cc2c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.416634] env[62235]: DEBUG nova.compute.manager [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Refreshing instance network info cache due to event network-changed-5a8cdc37-d4e8-4790-b051-82270896cc2c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 593.416784] env[62235]: DEBUG oslo_concurrency.lockutils [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] Acquiring lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.755143] env[62235]: DEBUG nova.compute.utils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.758438] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.758768] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 593.832379] env[62235]: DEBUG nova.policy [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3cc1724994843a19ceba8d8d10a6233', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c67f85102c2043118ea564c35be9dc68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.912505] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Releasing lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.912505] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.912505] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.912505] env[62235]: DEBUG oslo_concurrency.lockutils [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] Acquired lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.912505] env[62235]: DEBUG nova.network.neutron [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Refreshing network info cache for port 5a8cdc37-d4e8-4790-b051-82270896cc2c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 593.913474] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-911a09d5-8053-46a9-aacc-c5152e8d1f25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.929574] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95dd5383-adcf-4128-9369-b9c12eb3ea5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.954273] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8efd1fa-0e33-423c-975e-b2583e1cfbd4 could not be found. [ 593.954533] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.954718] env[62235]: INFO nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.954972] env[62235]: DEBUG oslo.service.loopingcall [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.955191] env[62235]: DEBUG nova.compute.manager [-] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.955285] env[62235]: DEBUG nova.network.neutron [-] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.976678] env[62235]: DEBUG nova.network.neutron [-] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.234515] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Successfully created port: 96ab7d45-c82f-43ab-ab41-216dd0378e09 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.258832] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 594.449842] env[62235]: DEBUG nova.network.neutron [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.478193] env[62235]: DEBUG nova.network.neutron [-] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.604397] env[62235]: DEBUG nova.network.neutron [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.754290] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d816d3-2e16-4a74-b83d-6e561536603d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.762372] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85c052f-d738-403d-aaa7-00a5cb14ab2a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.803193] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622d9a71-ea2d-4083-ac97-d97563f69f73 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.810497] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6544ed-907a-4fc8-82b0-f2fa0952a8ea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.824644] env[62235]: DEBUG nova.compute.provider_tree [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.980736] env[62235]: INFO nova.compute.manager [-] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Took 1.03 seconds to deallocate network for instance. [ 594.985148] env[62235]: DEBUG nova.compute.claims [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.985148] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.111630] env[62235]: DEBUG oslo_concurrency.lockutils [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] Releasing lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.111971] env[62235]: DEBUG nova.compute.manager [req-61bd3f4f-9fc3-4ba0-a15e-8d9e5494b7e2 req-b0605726-6ac8-480b-9504-bc21df992803 service nova] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Received event network-vif-deleted-5a8cdc37-d4e8-4790-b051-82270896cc2c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.273414] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.283564] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Acquiring lock "ea4a3f54-388e-451f-9318-78943a169213" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.284077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Lock "ea4a3f54-388e-451f-9318-78943a169213" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.300518] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.300796] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.302017] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.302017] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.302017] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.302017] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.302017] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.302324] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.303620] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.303796] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.303970] env[62235]: DEBUG nova.virt.hardware [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.305067] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a32766-bbce-4bcf-b03e-33f7e7b9d3ea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.320065] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16c51d3-fb6b-447a-a08b-069876983162 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.328548] env[62235]: DEBUG nova.scheduler.client.report [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.789355] env[62235]: DEBUG nova.compute.manager [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Received event network-changed-96ab7d45-c82f-43ab-ab41-216dd0378e09 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.789604] env[62235]: DEBUG nova.compute.manager [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Refreshing instance network info cache due to event network-changed-96ab7d45-c82f-43ab-ab41-216dd0378e09. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 595.789848] env[62235]: DEBUG oslo_concurrency.lockutils [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] Acquiring lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.790174] env[62235]: DEBUG oslo_concurrency.lockutils [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] Acquired lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.790417] env[62235]: DEBUG nova.network.neutron [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Refreshing network info cache for port 96ab7d45-c82f-43ab-ab41-216dd0378e09 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.843915] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.844453] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.846983] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.063s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.062853] env[62235]: ERROR nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. [ 596.062853] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.062853] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.062853] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.062853] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.062853] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.062853] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.062853] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.062853] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.062853] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 596.062853] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.062853] env[62235]: ERROR nova.compute.manager raise self.value [ 596.062853] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.062853] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.062853] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.062853] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.063818] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.063818] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.063818] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. [ 596.063818] env[62235]: ERROR nova.compute.manager [ 596.063818] env[62235]: Traceback (most recent call last): [ 596.063818] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.063818] env[62235]: listener.cb(fileno) [ 596.063818] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.063818] env[62235]: result = function(*args, **kwargs) [ 596.063818] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.063818] env[62235]: return func(*args, **kwargs) [ 596.063818] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.063818] env[62235]: raise e [ 596.063818] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.063818] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 596.063818] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.063818] env[62235]: created_port_ids = self._update_ports_for_instance( [ 596.063818] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.063818] env[62235]: with excutils.save_and_reraise_exception(): [ 596.063818] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.063818] env[62235]: self.force_reraise() [ 596.063818] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.063818] env[62235]: raise self.value [ 596.063818] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.063818] env[62235]: updated_port = self._update_port( [ 596.063818] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.063818] env[62235]: _ensure_no_port_binding_failure(port) [ 596.063818] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.063818] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.091865] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. [ 596.091865] env[62235]: Removing descriptor: 22 [ 596.091865] env[62235]: ERROR nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Traceback (most recent call last): [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] yield resources [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self.driver.spawn(context, instance, image_meta, [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.091865] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] vm_ref = self.build_virtual_machine(instance, [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] for vif in network_info: [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return self._sync_wrapper(fn, *args, **kwargs) [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self.wait() [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self[:] = self._gt.wait() [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return self._exit_event.wait() [ 596.092792] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] result = hub.switch() [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return self.greenlet.switch() [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] result = function(*args, **kwargs) [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return func(*args, **kwargs) [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] raise e [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] nwinfo = self.network_api.allocate_for_instance( [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.093372] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] created_port_ids = self._update_ports_for_instance( [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] with excutils.save_and_reraise_exception(): [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self.force_reraise() [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] raise self.value [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] updated_port = self._update_port( [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] _ensure_no_port_binding_failure(port) [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.093872] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] raise exception.PortBindingFailed(port_id=port['id']) [ 596.094113] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] nova.exception.PortBindingFailed: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. [ 596.094113] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] [ 596.094113] env[62235]: INFO nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Terminating instance [ 596.094113] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Acquiring lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.314113] env[62235]: DEBUG nova.network.neutron [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.351451] env[62235]: DEBUG nova.compute.utils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.352834] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.353519] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 596.375985] env[62235]: DEBUG nova.network.neutron [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.445608] env[62235]: DEBUG nova.policy [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbff33ad47a748b8aec74685b232ff2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b5362773ed741649896bd8296c0937d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.813733] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Successfully created port: 5be31e87-cd26-4ec7-a9a6-20b18809bdb1 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.832173] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca679a5-19b9-4287-b1c2-ae749e35d59f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.840027] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a7934a-2e61-4ab7-bbb9-fe15a9253c16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.873653] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.877124] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bab0a46-e938-43f7-91b6-2b855457e1ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.880267] env[62235]: DEBUG oslo_concurrency.lockutils [req-d324608d-8e9e-42f8-86ec-8a79c324df30 req-eb187edf-6444-44e4-bacc-4b38d2d0352e service nova] Releasing lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.880726] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Acquired lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.881084] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.889348] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaac53a3-c6ec-4f34-9bc0-746ad41d029d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.906335] env[62235]: DEBUG nova.compute.provider_tree [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.408752] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.411076] env[62235]: DEBUG nova.scheduler.client.report [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.521751] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.888837] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.916077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.916706] env[62235]: ERROR nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Traceback (most recent call last): [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self.driver.spawn(context, instance, image_meta, [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] vm_ref = self.build_virtual_machine(instance, [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.916706] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] for vif in network_info: [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return self._sync_wrapper(fn, *args, **kwargs) [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self.wait() [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self[:] = self._gt.wait() [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return self._exit_event.wait() [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] result = hub.switch() [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.916988] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return self.greenlet.switch() [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] result = function(*args, **kwargs) [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] return func(*args, **kwargs) [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] raise e [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] nwinfo = self.network_api.allocate_for_instance( [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] created_port_ids = self._update_ports_for_instance( [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] with excutils.save_and_reraise_exception(): [ 597.917303] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] self.force_reraise() [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] raise self.value [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] updated_port = self._update_port( [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] _ensure_no_port_binding_failure(port) [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] raise exception.PortBindingFailed(port_id=port['id']) [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] nova.exception.PortBindingFailed: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. [ 597.917681] env[62235]: ERROR nova.compute.manager [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] [ 597.918025] env[62235]: DEBUG nova.compute.utils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.920429] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.920646] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.920830] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.921041] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.921192] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.921336] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.921539] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.921722] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.921907] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.922200] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.922401] env[62235]: DEBUG nova.virt.hardware [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.922709] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.387s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.925766] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899ac6c7-e950-4f53-b424-71f56de01591 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.928586] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Build of instance 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf was re-scheduled: Binding failed for port f1a29580-7286-4eab-bbcf-0149f95ed7b2, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.929027] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.929258] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Acquiring lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.929407] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Acquired lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.929563] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.936243] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cac2b54-1a7e-4de1-8710-19904946eda3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.025440] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Releasing lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.025894] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.026154] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.026455] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-287f401c-c6f9-4e0e-8a99-802b973842ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.035562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50772ed-11d5-41cf-a3d4-fdfbab8a251a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.050517] env[62235]: DEBUG nova.compute.manager [req-fd639141-ddc9-488a-a157-15fcadd5aed7 req-d9c5b90f-ef99-43ba-8270-f046a3a27489 service nova] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Received event network-vif-deleted-96ab7d45-c82f-43ab-ab41-216dd0378e09 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.058707] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 028e94f4-6b8f-45c3-a046-ed44e575362a could not be found. [ 598.058918] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 598.059112] env[62235]: INFO nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 598.059435] env[62235]: DEBUG oslo.service.loopingcall [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.059645] env[62235]: DEBUG nova.compute.manager [-] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.059767] env[62235]: DEBUG nova.network.neutron [-] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.089577] env[62235]: DEBUG nova.network.neutron [-] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.333313] env[62235]: ERROR nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. [ 598.333313] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.333313] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.333313] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.333313] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.333313] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.333313] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.333313] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.333313] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.333313] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 598.333313] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.333313] env[62235]: ERROR nova.compute.manager raise self.value [ 598.333313] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.333313] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.333313] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.333313] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.333649] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.333649] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.333649] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. [ 598.333649] env[62235]: ERROR nova.compute.manager [ 598.333649] env[62235]: Traceback (most recent call last): [ 598.333649] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.333649] env[62235]: listener.cb(fileno) [ 598.333649] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.333649] env[62235]: result = function(*args, **kwargs) [ 598.333649] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.333649] env[62235]: return func(*args, **kwargs) [ 598.333649] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.333649] env[62235]: raise e [ 598.333649] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.333649] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 598.333649] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.333649] env[62235]: created_port_ids = self._update_ports_for_instance( [ 598.333649] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.333649] env[62235]: with excutils.save_and_reraise_exception(): [ 598.333649] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.333649] env[62235]: self.force_reraise() [ 598.333649] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.333649] env[62235]: raise self.value [ 598.333649] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.333649] env[62235]: updated_port = self._update_port( [ 598.333649] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.333649] env[62235]: _ensure_no_port_binding_failure(port) [ 598.333649] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.333649] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.334241] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. [ 598.334241] env[62235]: Removing descriptor: 16 [ 598.334241] env[62235]: ERROR nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Traceback (most recent call last): [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] yield resources [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self.driver.spawn(context, instance, image_meta, [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.334241] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] vm_ref = self.build_virtual_machine(instance, [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] for vif in network_info: [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return self._sync_wrapper(fn, *args, **kwargs) [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self.wait() [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self[:] = self._gt.wait() [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return self._exit_event.wait() [ 598.334476] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] result = hub.switch() [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return self.greenlet.switch() [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] result = function(*args, **kwargs) [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return func(*args, **kwargs) [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] raise e [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] nwinfo = self.network_api.allocate_for_instance( [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.334724] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] created_port_ids = self._update_ports_for_instance( [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] with excutils.save_and_reraise_exception(): [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self.force_reraise() [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] raise self.value [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] updated_port = self._update_port( [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] _ensure_no_port_binding_failure(port) [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.334972] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] raise exception.PortBindingFailed(port_id=port['id']) [ 598.335214] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] nova.exception.PortBindingFailed: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. [ 598.335214] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] [ 598.335214] env[62235]: INFO nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Terminating instance [ 598.337609] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Acquiring lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.337808] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Acquired lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.338123] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.451088] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.544464] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.591215] env[62235]: DEBUG nova.network.neutron [-] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.824166] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc42e1de-0fc6-4e13-8e02-78b0fcac6fbf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.831805] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d366d3-7394-45e3-9309-621a77385a65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.864308] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8147166-d1f6-4b76-be45-0557769670d0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.871635] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd0c976-5594-4f3c-ab58-584c36347555 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.885916] env[62235]: DEBUG nova.compute.provider_tree [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.888750] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.990603] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.047619] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Releasing lock "refresh_cache-1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.047952] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.049785] env[62235]: DEBUG nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.049785] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.071399] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.098249] env[62235]: INFO nova.compute.manager [-] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Took 1.04 seconds to deallocate network for instance. [ 599.098820] env[62235]: DEBUG nova.compute.claims [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.098981] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.390049] env[62235]: DEBUG nova.scheduler.client.report [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.493756] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Releasing lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.494206] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.494404] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.494694] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d97c2243-bab4-4f09-8e31-cd3c151f1072 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.503027] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3872807-c6a8-4f50-bc4c-d16c4654ab72 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.525033] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 41adaeb9-c39f-4214-b9d1-7137aab3455f could not be found. [ 599.525201] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.525374] env[62235]: INFO nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 599.525613] env[62235]: DEBUG oslo.service.loopingcall [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.525838] env[62235]: DEBUG nova.compute.manager [-] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.525931] env[62235]: DEBUG nova.network.neutron [-] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.539021] env[62235]: DEBUG nova.network.neutron [-] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.574321] env[62235]: DEBUG nova.network.neutron [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.894339] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.971s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.894822] env[62235]: ERROR nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] Traceback (most recent call last): [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self.driver.spawn(context, instance, image_meta, [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] vm_ref = self.build_virtual_machine(instance, [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.894822] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] for vif in network_info: [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return self._sync_wrapper(fn, *args, **kwargs) [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self.wait() [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self[:] = self._gt.wait() [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return self._exit_event.wait() [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] result = hub.switch() [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.895067] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return self.greenlet.switch() [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] result = function(*args, **kwargs) [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] return func(*args, **kwargs) [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] raise e [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] nwinfo = self.network_api.allocate_for_instance( [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] created_port_ids = self._update_ports_for_instance( [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] with excutils.save_and_reraise_exception(): [ 599.895325] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] self.force_reraise() [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] raise self.value [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] updated_port = self._update_port( [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] _ensure_no_port_binding_failure(port) [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] raise exception.PortBindingFailed(port_id=port['id']) [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] nova.exception.PortBindingFailed: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. [ 599.895631] env[62235]: ERROR nova.compute.manager [instance: 842f0697-6cec-423b-a35a-583985713a08] [ 599.895871] env[62235]: DEBUG nova.compute.utils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.897098] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.008s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.898536] env[62235]: INFO nova.compute.claims [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.901266] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Build of instance 842f0697-6cec-423b-a35a-583985713a08 was re-scheduled: Binding failed for port 506492c3-fdb0-4d15-afef-31713c721d2b, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.901702] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.901941] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Acquiring lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.902159] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Acquired lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.902344] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.041297] env[62235]: DEBUG nova.network.neutron [-] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.077274] env[62235]: INFO nova.compute.manager [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] [instance: 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf] Took 1.03 seconds to deallocate network for instance. [ 600.138263] env[62235]: DEBUG nova.compute.manager [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Received event network-changed-5be31e87-cd26-4ec7-a9a6-20b18809bdb1 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.138455] env[62235]: DEBUG nova.compute.manager [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Refreshing instance network info cache due to event network-changed-5be31e87-cd26-4ec7-a9a6-20b18809bdb1. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.138667] env[62235]: DEBUG oslo_concurrency.lockutils [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] Acquiring lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.138808] env[62235]: DEBUG oslo_concurrency.lockutils [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] Acquired lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.138966] env[62235]: DEBUG nova.network.neutron [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Refreshing network info cache for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 600.429734] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.514465] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.543832] env[62235]: INFO nova.compute.manager [-] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Took 1.02 seconds to deallocate network for instance. [ 600.546138] env[62235]: DEBUG nova.compute.claims [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.546864] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.661379] env[62235]: DEBUG nova.network.neutron [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.747740] env[62235]: DEBUG nova.network.neutron [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.022593] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Releasing lock "refresh_cache-842f0697-6cec-423b-a35a-583985713a08" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.022918] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.023140] env[62235]: DEBUG nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.023315] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.041724] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.109871] env[62235]: INFO nova.scheduler.client.report [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Deleted allocations for instance 1fdb958b-7a9e-4d15-9ff1-684a3412dcaf [ 601.251108] env[62235]: DEBUG oslo_concurrency.lockutils [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] Releasing lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.251371] env[62235]: DEBUG nova.compute.manager [req-7748a4a7-d0ea-4bca-967d-8f1248a362f5 req-94db9c49-0064-40b7-ad51-058186b4d939 service nova] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Received event network-vif-deleted-5be31e87-cd26-4ec7-a9a6-20b18809bdb1 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.394370] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a2c507-8328-485e-9fb8-88f06495dcd7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.402144] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d14abe8-dc9d-4ac6-b03c-e5ebd8ddbd73 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.436243] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999a9b13-4d5f-43e1-8df9-3411389b2211 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.443695] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8820bc-6e29-4cbd-8e30-9da5a8339196 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.456807] env[62235]: DEBUG nova.compute.provider_tree [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.544681] env[62235]: DEBUG nova.network.neutron [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.622669] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b9e787d3-df3a-4686-b39f-22d9cbee979d tempest-VolumesAssistedSnapshotsTest-1803393474 tempest-VolumesAssistedSnapshotsTest-1803393474-project-member] Lock "1fdb958b-7a9e-4d15-9ff1-684a3412dcaf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.182s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.960306] env[62235]: DEBUG nova.scheduler.client.report [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.047717] env[62235]: INFO nova.compute.manager [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] [instance: 842f0697-6cec-423b-a35a-583985713a08] Took 1.02 seconds to deallocate network for instance. [ 602.125697] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.467737] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.468308] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.471583] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.874s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.472452] env[62235]: INFO nova.compute.claims [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.652397] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.977182] env[62235]: DEBUG nova.compute.utils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.982244] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.982244] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 603.056283] env[62235]: DEBUG nova.policy [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82372abe3ded4f07aabcd06a89a9aa70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ac79e027d234eb1be8e023f6b2439b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.084488] env[62235]: INFO nova.scheduler.client.report [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Deleted allocations for instance 842f0697-6cec-423b-a35a-583985713a08 [ 603.411032] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Successfully created port: 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.482286] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.596963] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30cd98a3-c45d-4854-ab03-44bf7ec70e09 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147 tempest-FloatingIPsAssociationNegativeTestJSON-1285654147-project-member] Lock "842f0697-6cec-423b-a35a-583985713a08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.699s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.966581] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12119f6-9560-4a89-8a74-e2bc9af26e98 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.974970] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d83378-0d0d-41d9-928a-eca6c2771b55 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.015018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728f7209-e047-4056-8e09-56a4ff9f2286 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.020766] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90e4efc-dab1-44e1-9f3f-674b5ad5bf2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.038486] env[62235]: DEBUG nova.compute.provider_tree [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.099307] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 604.517290] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.546204] env[62235]: DEBUG nova.scheduler.client.report [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.550632] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.551038] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.551404] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.551750] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.552021] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.552263] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.552727] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.553788] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.553788] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.553788] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.553788] env[62235]: DEBUG nova.virt.hardware [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.554448] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c373a1ea-0c8a-4d43-ad06-0eb58b1db677 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.566582] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fa1f95-f153-402b-beba-45c79dea0a64 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.628404] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.943355] env[62235]: DEBUG nova.compute.manager [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Received event network-changed-3a14258d-df5f-4a8b-8ffe-5fb1f19b024d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.943546] env[62235]: DEBUG nova.compute.manager [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Refreshing instance network info cache due to event network-changed-3a14258d-df5f-4a8b-8ffe-5fb1f19b024d. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.943757] env[62235]: DEBUG oslo_concurrency.lockutils [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] Acquiring lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.943901] env[62235]: DEBUG oslo_concurrency.lockutils [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] Acquired lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.944271] env[62235]: DEBUG nova.network.neutron [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Refreshing network info cache for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 605.061449] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.062040] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 605.065353] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.226s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.066870] env[62235]: INFO nova.compute.claims [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.273557] env[62235]: ERROR nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. [ 605.273557] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.273557] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.273557] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.273557] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.273557] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.273557] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.273557] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.273557] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.273557] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 605.273557] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.273557] env[62235]: ERROR nova.compute.manager raise self.value [ 605.273557] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.273557] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.273557] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.273557] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.274160] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.274160] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.274160] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. [ 605.274160] env[62235]: ERROR nova.compute.manager [ 605.274160] env[62235]: Traceback (most recent call last): [ 605.274160] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.274160] env[62235]: listener.cb(fileno) [ 605.274160] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.274160] env[62235]: result = function(*args, **kwargs) [ 605.274160] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.274160] env[62235]: return func(*args, **kwargs) [ 605.274160] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.274160] env[62235]: raise e [ 605.274160] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.274160] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 605.274160] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.274160] env[62235]: created_port_ids = self._update_ports_for_instance( [ 605.274160] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.274160] env[62235]: with excutils.save_and_reraise_exception(): [ 605.274160] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.274160] env[62235]: self.force_reraise() [ 605.274160] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.274160] env[62235]: raise self.value [ 605.274160] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.274160] env[62235]: updated_port = self._update_port( [ 605.274160] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.274160] env[62235]: _ensure_no_port_binding_failure(port) [ 605.274160] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.274160] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.274847] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. [ 605.274847] env[62235]: Removing descriptor: 16 [ 605.274847] env[62235]: ERROR nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Traceback (most recent call last): [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] yield resources [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self.driver.spawn(context, instance, image_meta, [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.274847] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] vm_ref = self.build_virtual_machine(instance, [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] for vif in network_info: [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return self._sync_wrapper(fn, *args, **kwargs) [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self.wait() [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self[:] = self._gt.wait() [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return self._exit_event.wait() [ 605.275127] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] result = hub.switch() [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return self.greenlet.switch() [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] result = function(*args, **kwargs) [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return func(*args, **kwargs) [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] raise e [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] nwinfo = self.network_api.allocate_for_instance( [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.275413] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] created_port_ids = self._update_ports_for_instance( [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] with excutils.save_and_reraise_exception(): [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self.force_reraise() [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] raise self.value [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] updated_port = self._update_port( [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] _ensure_no_port_binding_failure(port) [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.275680] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] raise exception.PortBindingFailed(port_id=port['id']) [ 605.275939] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] nova.exception.PortBindingFailed: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. [ 605.275939] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] [ 605.275939] env[62235]: INFO nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Terminating instance [ 605.276729] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Acquiring lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.483799] env[62235]: DEBUG nova.network.neutron [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.568027] env[62235]: DEBUG nova.network.neutron [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.569930] env[62235]: DEBUG nova.compute.utils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.572091] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.572314] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 605.634640] env[62235]: DEBUG nova.policy [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '52543cbd68814006817043e7a2264aae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b72b4f3724234ed6b7315ea804fda0d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 606.012456] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Successfully created port: 09157d54-9213-410e-b1cb-8b7520390e36 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.079348] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.083336] env[62235]: DEBUG oslo_concurrency.lockutils [req-23c63943-e4af-4fbf-a23e-839b1083d0f8 req-83511121-d2d7-44bb-8e5d-00b670401273 service nova] Releasing lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.083975] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Acquired lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.084184] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.577008] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb1afa4-7826-44f6-b0da-b75e70e160fa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.591151] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137a656a-9f09-4040-a4bc-1978c4d53883 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.630383] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.633278] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01298496-3a2d-4518-9ea5-79cb27a71e67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.644476] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fefc8b6-7f58-46e8-a770-3abbef59c9a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.659304] env[62235]: DEBUG nova.compute.provider_tree [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.763340] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.098376] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 607.131953] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 607.132636] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 607.132994] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.133300] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 607.133516] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.133678] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 607.133893] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 607.134081] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 607.134460] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 607.134642] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 607.134820] env[62235]: DEBUG nova.virt.hardware [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.135772] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1dda35e-ecf1-4ab0-a6ad-ab7b4a60942b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.144546] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ac6313-4d26-42d0-93f6-c5453019e1c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.162294] env[62235]: DEBUG nova.scheduler.client.report [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.255810] env[62235]: DEBUG nova.compute.manager [req-ea50a397-46b1-4f2e-91b1-6163bbdb49c4 req-ba18393e-355b-4ca1-9ad4-be87d8fcec45 service nova] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Received event network-vif-deleted-3a14258d-df5f-4a8b-8ffe-5fb1f19b024d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.265150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Releasing lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.265650] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.265846] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.266131] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be43a054-a669-4486-acfe-e996e296e36c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.275023] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d10edca-1e67-4fe9-9ab5-0d4f1ab96fe9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.299682] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7684530-e42d-4a30-9c7c-28be8ff31b7b could not be found. [ 607.299682] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.299800] env[62235]: INFO nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 607.300047] env[62235]: DEBUG oslo.service.loopingcall [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.300253] env[62235]: DEBUG nova.compute.manager [-] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.300340] env[62235]: DEBUG nova.network.neutron [-] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.347880] env[62235]: DEBUG nova.network.neutron [-] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.667534] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.668086] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.670919] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.025s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.850924] env[62235]: DEBUG nova.network.neutron [-] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.078295] env[62235]: ERROR nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. [ 608.078295] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.078295] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.078295] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.078295] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.078295] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.078295] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.078295] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.078295] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.078295] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 608.078295] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.078295] env[62235]: ERROR nova.compute.manager raise self.value [ 608.078295] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.078295] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.078295] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.078295] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.079095] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.079095] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.079095] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. [ 608.079095] env[62235]: ERROR nova.compute.manager [ 608.079095] env[62235]: Traceback (most recent call last): [ 608.079095] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.079095] env[62235]: listener.cb(fileno) [ 608.079095] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.079095] env[62235]: result = function(*args, **kwargs) [ 608.079095] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.079095] env[62235]: return func(*args, **kwargs) [ 608.079095] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.079095] env[62235]: raise e [ 608.079095] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.079095] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 608.079095] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.079095] env[62235]: created_port_ids = self._update_ports_for_instance( [ 608.079095] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.079095] env[62235]: with excutils.save_and_reraise_exception(): [ 608.079095] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.079095] env[62235]: self.force_reraise() [ 608.079095] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.079095] env[62235]: raise self.value [ 608.079095] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.079095] env[62235]: updated_port = self._update_port( [ 608.079095] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.079095] env[62235]: _ensure_no_port_binding_failure(port) [ 608.079095] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.079095] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.079958] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. [ 608.079958] env[62235]: Removing descriptor: 16 [ 608.079958] env[62235]: ERROR nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Traceback (most recent call last): [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] yield resources [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self.driver.spawn(context, instance, image_meta, [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.079958] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] vm_ref = self.build_virtual_machine(instance, [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] for vif in network_info: [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return self._sync_wrapper(fn, *args, **kwargs) [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self.wait() [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self[:] = self._gt.wait() [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return self._exit_event.wait() [ 608.080299] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] result = hub.switch() [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return self.greenlet.switch() [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] result = function(*args, **kwargs) [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return func(*args, **kwargs) [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] raise e [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] nwinfo = self.network_api.allocate_for_instance( [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.080627] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] created_port_ids = self._update_ports_for_instance( [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] with excutils.save_and_reraise_exception(): [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self.force_reraise() [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] raise self.value [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] updated_port = self._update_port( [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] _ensure_no_port_binding_failure(port) [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.081050] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] raise exception.PortBindingFailed(port_id=port['id']) [ 608.081365] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] nova.exception.PortBindingFailed: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. [ 608.081365] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] [ 608.081365] env[62235]: INFO nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Terminating instance [ 608.085274] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Acquiring lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.085274] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Acquired lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.085274] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.178339] env[62235]: DEBUG nova.compute.utils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.187824] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.187824] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 608.267837] env[62235]: DEBUG nova.policy [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19727a6117df4d298b926f707483a8e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2171f5abf1ef4e1987b9a77d132215bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.356844] env[62235]: INFO nova.compute.manager [-] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Took 1.06 seconds to deallocate network for instance. [ 608.363236] env[62235]: DEBUG nova.compute.claims [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.363469] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.635340] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.690462] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf11888d-5106-4e0a-b7fb-9e7e540b7320 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.695069] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.703646] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1216fd-d99d-4730-88b9-58a6c7fcaeed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.747990] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048bc426-0071-4572-b146-56d8a2fb19aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.758989] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c32598-aa5a-4aad-a241-941d11f090da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.775662] env[62235]: DEBUG nova.compute.provider_tree [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.850704] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.915879] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Successfully created port: 1dfea62d-9378-4e4e-99aa-46c49f11a12d {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.279759] env[62235]: DEBUG nova.scheduler.client.report [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.287895] env[62235]: DEBUG nova.compute.manager [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Received event network-changed-09157d54-9213-410e-b1cb-8b7520390e36 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.287895] env[62235]: DEBUG nova.compute.manager [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Refreshing instance network info cache due to event network-changed-09157d54-9213-410e-b1cb-8b7520390e36. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 609.287895] env[62235]: DEBUG oslo_concurrency.lockutils [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] Acquiring lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.354395] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Releasing lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.355110] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.355318] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.355641] env[62235]: DEBUG oslo_concurrency.lockutils [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] Acquired lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.355818] env[62235]: DEBUG nova.network.neutron [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Refreshing network info cache for port 09157d54-9213-410e-b1cb-8b7520390e36 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 609.357618] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00c723d7-2f19-4651-ab2a-d6f613669961 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.371490] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbd01e2-4a3f-41f4-92aa-75845f3cbb9b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.400667] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ec00f2ba-ed0c-47de-a9b2-bd277f96e65f could not be found. [ 609.400898] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 609.402036] env[62235]: INFO nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 609.402036] env[62235]: DEBUG oslo.service.loopingcall [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.402036] env[62235]: DEBUG nova.compute.manager [-] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.402036] env[62235]: DEBUG nova.network.neutron [-] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.427587] env[62235]: DEBUG nova.network.neutron [-] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.712267] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.742790] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.742934] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.743103] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.743292] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.743437] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.743582] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.743793] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.743946] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.744122] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.744282] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.744449] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.745313] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11923b3-f719-4f15-87d2-262c5adafff0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.755490] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744bcf6f-fa4c-49ed-b87e-328a3e3a1e3e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.786294] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.115s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.786921] env[62235]: ERROR nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Traceback (most recent call last): [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self.driver.spawn(context, instance, image_meta, [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] vm_ref = self.build_virtual_machine(instance, [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.786921] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] for vif in network_info: [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] return self._sync_wrapper(fn, *args, **kwargs) [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self.wait() [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self[:] = self._gt.wait() [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] return self._exit_event.wait() [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] current.throw(*self._exc) [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.787215] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] result = function(*args, **kwargs) [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] return func(*args, **kwargs) [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] raise e [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] nwinfo = self.network_api.allocate_for_instance( [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] created_port_ids = self._update_ports_for_instance( [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] with excutils.save_and_reraise_exception(): [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] self.force_reraise() [ 609.787671] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] raise self.value [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] updated_port = self._update_port( [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] _ensure_no_port_binding_failure(port) [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] raise exception.PortBindingFailed(port_id=port['id']) [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] nova.exception.PortBindingFailed: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. [ 609.787976] env[62235]: ERROR nova.compute.manager [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] [ 609.787976] env[62235]: DEBUG nova.compute.utils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.790618] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.408s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.793762] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Build of instance ab123e55-fa0e-4216-9afc-f8672598b08e was re-scheduled: Binding failed for port 064a1463-1523-4a54-9bd7-c0e5d15800e0, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.794212] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.794435] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.794582] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquired lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.794739] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.886184] env[62235]: DEBUG nova.network.neutron [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.931042] env[62235]: DEBUG nova.network.neutron [-] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.039712] env[62235]: DEBUG nova.network.neutron [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.137416] env[62235]: ERROR nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. [ 610.137416] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.137416] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.137416] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.137416] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.137416] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.137416] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.137416] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.137416] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.137416] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 610.137416] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.137416] env[62235]: ERROR nova.compute.manager raise self.value [ 610.137416] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.137416] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.137416] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.137416] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.137923] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.137923] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.137923] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. [ 610.137923] env[62235]: ERROR nova.compute.manager [ 610.137923] env[62235]: Traceback (most recent call last): [ 610.137923] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.137923] env[62235]: listener.cb(fileno) [ 610.137923] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.137923] env[62235]: result = function(*args, **kwargs) [ 610.137923] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.137923] env[62235]: return func(*args, **kwargs) [ 610.137923] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.137923] env[62235]: raise e [ 610.137923] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.137923] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 610.137923] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.137923] env[62235]: created_port_ids = self._update_ports_for_instance( [ 610.137923] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.137923] env[62235]: with excutils.save_and_reraise_exception(): [ 610.137923] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.137923] env[62235]: self.force_reraise() [ 610.137923] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.137923] env[62235]: raise self.value [ 610.137923] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.137923] env[62235]: updated_port = self._update_port( [ 610.137923] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.137923] env[62235]: _ensure_no_port_binding_failure(port) [ 610.137923] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.137923] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.138624] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. [ 610.138624] env[62235]: Removing descriptor: 16 [ 610.140779] env[62235]: ERROR nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Traceback (most recent call last): [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] yield resources [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self.driver.spawn(context, instance, image_meta, [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] vm_ref = self.build_virtual_machine(instance, [ 610.140779] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] for vif in network_info: [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return self._sync_wrapper(fn, *args, **kwargs) [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self.wait() [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self[:] = self._gt.wait() [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return self._exit_event.wait() [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.141132] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] result = hub.switch() [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return self.greenlet.switch() [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] result = function(*args, **kwargs) [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return func(*args, **kwargs) [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] raise e [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] nwinfo = self.network_api.allocate_for_instance( [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] created_port_ids = self._update_ports_for_instance( [ 610.141404] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] with excutils.save_and_reraise_exception(): [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self.force_reraise() [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] raise self.value [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] updated_port = self._update_port( [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] _ensure_no_port_binding_failure(port) [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] raise exception.PortBindingFailed(port_id=port['id']) [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] nova.exception.PortBindingFailed: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. [ 610.141750] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] [ 610.142464] env[62235]: INFO nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Terminating instance [ 610.143777] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.145753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.145753] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.336961] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.423369] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.435176] env[62235]: INFO nova.compute.manager [-] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Took 1.03 seconds to deallocate network for instance. [ 610.438215] env[62235]: DEBUG nova.compute.claims [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.438829] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.543308] env[62235]: DEBUG oslo_concurrency.lockutils [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] Releasing lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.543571] env[62235]: DEBUG nova.compute.manager [req-afa0cad3-e522-4a16-81db-9b1689f3c42e req-5c8ae2fd-14fb-4b5e-8c9a-3e2086165f40 service nova] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Received event network-vif-deleted-09157d54-9213-410e-b1cb-8b7520390e36 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.689078] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.708689] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Acquiring lock "3b4167dd-4535-4482-9452-5548e07dc581" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.708944] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Lock "3b4167dd-4535-4482-9452-5548e07dc581" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.800228] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.807449] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f0cacb-0cf1-435f-8ce8-5d831eda467a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.816694] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd33c8c-3e94-4bdd-bf45-e2e955b4aa28 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.850738] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb52baf9-787d-4f16-877b-00837d6bdd6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.858585] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d785ea66-d0f7-40cd-a2d8-8c9833b7592e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.872515] env[62235]: DEBUG nova.compute.provider_tree [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.926142] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Releasing lock "refresh_cache-ab123e55-fa0e-4216-9afc-f8672598b08e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.926387] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.926579] env[62235]: DEBUG nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.926776] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.941543] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.305697] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.306137] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.306307] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.306619] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb5544d9-cfc5-4955-b4bb-409276324408 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.323152] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dad7dfb-ee63-4529-a412-4cbf2c6a6422 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.347555] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a84e793b-4dbd-4b89-8fc4-94baacecd8e7 could not be found. [ 611.347854] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.348109] env[62235]: INFO nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 611.348415] env[62235]: DEBUG oslo.service.loopingcall [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.351474] env[62235]: DEBUG nova.compute.manager [-] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.351474] env[62235]: DEBUG nova.network.neutron [-] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.373470] env[62235]: DEBUG nova.network.neutron [-] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.375371] env[62235]: DEBUG nova.scheduler.client.report [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.400255] env[62235]: DEBUG nova.compute.manager [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Received event network-changed-1dfea62d-9378-4e4e-99aa-46c49f11a12d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.400462] env[62235]: DEBUG nova.compute.manager [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Refreshing instance network info cache due to event network-changed-1dfea62d-9378-4e4e-99aa-46c49f11a12d. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 611.400690] env[62235]: DEBUG oslo_concurrency.lockutils [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] Acquiring lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.400877] env[62235]: DEBUG oslo_concurrency.lockutils [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] Acquired lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.401059] env[62235]: DEBUG nova.network.neutron [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Refreshing network info cache for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.443809] env[62235]: DEBUG nova.network.neutron [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.881133] env[62235]: DEBUG nova.network.neutron [-] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.881816] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.091s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.882434] env[62235]: ERROR nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Traceback (most recent call last): [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self.driver.spawn(context, instance, image_meta, [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] vm_ref = self.build_virtual_machine(instance, [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.882434] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] for vif in network_info: [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] return self._sync_wrapper(fn, *args, **kwargs) [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self.wait() [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self[:] = self._gt.wait() [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] return self._exit_event.wait() [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] current.throw(*self._exc) [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.882774] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] result = function(*args, **kwargs) [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] return func(*args, **kwargs) [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] raise e [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] nwinfo = self.network_api.allocate_for_instance( [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] created_port_ids = self._update_ports_for_instance( [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] with excutils.save_and_reraise_exception(): [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] self.force_reraise() [ 611.883231] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] raise self.value [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] updated_port = self._update_port( [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] _ensure_no_port_binding_failure(port) [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] raise exception.PortBindingFailed(port_id=port['id']) [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] nova.exception.PortBindingFailed: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. [ 611.883596] env[62235]: ERROR nova.compute.manager [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] [ 611.887593] env[62235]: DEBUG nova.compute.utils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.887593] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.902s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.889598] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Build of instance b6122fc8-989d-4321-ac7a-80b2402a695f was re-scheduled: Binding failed for port 922b5a90-358d-4fee-8bcf-5087b4e9e1be, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.890337] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.890683] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Acquiring lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.890816] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Acquired lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.890989] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.932205] env[62235]: DEBUG nova.network.neutron [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.947184] env[62235]: INFO nova.compute.manager [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: ab123e55-fa0e-4216-9afc-f8672598b08e] Took 1.02 seconds to deallocate network for instance. [ 612.099058] env[62235]: DEBUG nova.network.neutron [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.387295] env[62235]: INFO nova.compute.manager [-] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Took 1.04 seconds to deallocate network for instance. [ 612.389950] env[62235]: DEBUG nova.compute.claims [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 612.390147] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.422043] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.593179] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.602538] env[62235]: DEBUG oslo_concurrency.lockutils [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] Releasing lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.602538] env[62235]: DEBUG nova.compute.manager [req-700fe841-2963-4353-89e8-8f54fb3a0836 req-fe75fb38-8d80-4b43-9997-72d2da94151c service nova] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Received event network-vif-deleted-1dfea62d-9378-4e4e-99aa-46c49f11a12d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.892110] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382c52be-e6f8-41f6-8978-4b140f361fb1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.898349] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d536f3-2028-4b77-ad45-9ba2316d452a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.930261] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2091a17c-ea26-4f65-96b4-d5f1e5877638 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.938526] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac41cfae-64b9-4bdb-8675-cc5d1e1b09eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.953805] env[62235]: DEBUG nova.compute.provider_tree [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.980674] env[62235]: INFO nova.scheduler.client.report [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Deleted allocations for instance ab123e55-fa0e-4216-9afc-f8672598b08e [ 613.094256] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Releasing lock "refresh_cache-b6122fc8-989d-4321-ac7a-80b2402a695f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.094496] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.094694] env[62235]: DEBUG nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.094893] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.116498] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.458591] env[62235]: DEBUG nova.scheduler.client.report [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.490959] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b72c7c7a-7eac-4e0d-ba4e-34006ff3909f tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "ab123e55-fa0e-4216-9afc-f8672598b08e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.355s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.620063] env[62235]: DEBUG nova.network.neutron [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.970329] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.080s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.970329] env[62235]: ERROR nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. [ 613.970329] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Traceback (most recent call last): [ 613.970329] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.970329] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self.driver.spawn(context, instance, image_meta, [ 613.970329] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.970329] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.970329] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.970329] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] vm_ref = self.build_virtual_machine(instance, [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] for vif in network_info: [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return self._sync_wrapper(fn, *args, **kwargs) [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self.wait() [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self[:] = self._gt.wait() [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return self._exit_event.wait() [ 613.970696] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] result = hub.switch() [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return self.greenlet.switch() [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] result = function(*args, **kwargs) [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] return func(*args, **kwargs) [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] raise e [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] nwinfo = self.network_api.allocate_for_instance( [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.970980] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] created_port_ids = self._update_ports_for_instance( [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] with excutils.save_and_reraise_exception(): [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] self.force_reraise() [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] raise self.value [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] updated_port = self._update_port( [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] _ensure_no_port_binding_failure(port) [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.971266] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] raise exception.PortBindingFailed(port_id=port['id']) [ 613.971613] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] nova.exception.PortBindingFailed: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. [ 613.971613] env[62235]: ERROR nova.compute.manager [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] [ 613.971613] env[62235]: DEBUG nova.compute.utils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.971613] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.871s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.975068] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Build of instance c8efd1fa-0e33-423c-975e-b2583e1cfbd4 was re-scheduled: Binding failed for port 5a8cdc37-d4e8-4790-b051-82270896cc2c, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.975534] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.975773] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquiring lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.975921] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Acquired lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.976091] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.994525] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.122396] env[62235]: INFO nova.compute.manager [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] [instance: b6122fc8-989d-4321-ac7a-80b2402a695f] Took 1.03 seconds to deallocate network for instance. [ 614.510648] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.525898] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.612427] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.919285] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5941aa56-29f5-4f13-ace2-2f9a701cd816 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.930690] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9db9e6-1857-468e-a73c-cc4d718bdae3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.966638] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20355c4c-97cc-4527-aaa9-f669abbbce58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.975221] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc217bc-4046-40d1-8209-f215285e17c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.988295] env[62235]: DEBUG nova.compute.provider_tree [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.115442] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Releasing lock "refresh_cache-c8efd1fa-0e33-423c-975e-b2583e1cfbd4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.115677] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.115869] env[62235]: DEBUG nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.116051] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.145422] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.155266] env[62235]: INFO nova.scheduler.client.report [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Deleted allocations for instance b6122fc8-989d-4321-ac7a-80b2402a695f [ 615.491361] env[62235]: DEBUG nova.scheduler.client.report [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.649268] env[62235]: DEBUG nova.network.neutron [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.668307] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3de62e99-ba8b-4748-a04d-3a2bd7e21053 tempest-ServersWithSpecificFlavorTestJSON-1407557860 tempest-ServersWithSpecificFlavorTestJSON-1407557860-project-member] Lock "b6122fc8-989d-4321-ac7a-80b2402a695f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.634s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.001036] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.001036] env[62235]: ERROR nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. [ 616.001036] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Traceback (most recent call last): [ 616.001036] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.001036] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self.driver.spawn(context, instance, image_meta, [ 616.001036] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.001036] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.001036] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.001036] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] vm_ref = self.build_virtual_machine(instance, [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] for vif in network_info: [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return self._sync_wrapper(fn, *args, **kwargs) [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self.wait() [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self[:] = self._gt.wait() [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return self._exit_event.wait() [ 616.001630] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] result = hub.switch() [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return self.greenlet.switch() [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] result = function(*args, **kwargs) [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] return func(*args, **kwargs) [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] raise e [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] nwinfo = self.network_api.allocate_for_instance( [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.001947] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] created_port_ids = self._update_ports_for_instance( [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] with excutils.save_and_reraise_exception(): [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] self.force_reraise() [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] raise self.value [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] updated_port = self._update_port( [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] _ensure_no_port_binding_failure(port) [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.002289] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] raise exception.PortBindingFailed(port_id=port['id']) [ 616.002546] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] nova.exception.PortBindingFailed: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. [ 616.002546] env[62235]: ERROR nova.compute.manager [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] [ 616.002546] env[62235]: DEBUG nova.compute.utils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.002546] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.455s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.007706] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Build of instance 028e94f4-6b8f-45c3-a046-ed44e575362a was re-scheduled: Binding failed for port 96ab7d45-c82f-43ab-ab41-216dd0378e09, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 616.009488] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 616.009488] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Acquiring lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.009488] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Acquired lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.009488] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.152040] env[62235]: INFO nova.compute.manager [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] [instance: c8efd1fa-0e33-423c-975e-b2583e1cfbd4] Took 1.04 seconds to deallocate network for instance. [ 616.170935] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.538348] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.594863] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.691403] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.939019] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd2e5fd-21ed-44f3-b249-4891faf99f3b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.946945] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6dfb3b-977c-43de-845d-4940b82d4b2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.977487] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d802b9-2988-4030-9d73-3f3ff7b8792a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.984989] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4806656c-0257-4f58-a12d-66e5d9f7b921 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.998763] env[62235]: DEBUG nova.compute.provider_tree [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.097620] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Releasing lock "refresh_cache-028e94f4-6b8f-45c3-a046-ed44e575362a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.097881] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 617.098238] env[62235]: DEBUG nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.098436] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 617.122656] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.195969] env[62235]: INFO nova.scheduler.client.report [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Deleted allocations for instance c8efd1fa-0e33-423c-975e-b2583e1cfbd4 [ 617.503445] env[62235]: DEBUG nova.scheduler.client.report [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.627504] env[62235]: DEBUG nova.network.neutron [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.707181] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a65b608-3425-4610-98ec-1d1b2161b386 tempest-ListImageFiltersTestJSON-31327926 tempest-ListImageFiltersTestJSON-31327926-project-member] Lock "c8efd1fa-0e33-423c-975e-b2583e1cfbd4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.018s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.824458] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "86e61430-ce9a-436e-9331-6276604610e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.824458] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "86e61430-ce9a-436e-9331-6276604610e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.010131] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.010131] env[62235]: ERROR nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. [ 618.010131] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Traceback (most recent call last): [ 618.010131] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.010131] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self.driver.spawn(context, instance, image_meta, [ 618.010131] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.010131] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.010131] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.010131] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] vm_ref = self.build_virtual_machine(instance, [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] for vif in network_info: [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return self._sync_wrapper(fn, *args, **kwargs) [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self.wait() [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self[:] = self._gt.wait() [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return self._exit_event.wait() [ 618.010418] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] result = hub.switch() [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return self.greenlet.switch() [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] result = function(*args, **kwargs) [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] return func(*args, **kwargs) [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] raise e [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] nwinfo = self.network_api.allocate_for_instance( [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.010710] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] created_port_ids = self._update_ports_for_instance( [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] with excutils.save_and_reraise_exception(): [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] self.force_reraise() [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] raise self.value [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] updated_port = self._update_port( [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] _ensure_no_port_binding_failure(port) [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.010997] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] raise exception.PortBindingFailed(port_id=port['id']) [ 618.011261] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] nova.exception.PortBindingFailed: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. [ 618.011261] env[62235]: ERROR nova.compute.manager [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] [ 618.011261] env[62235]: DEBUG nova.compute.utils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.015022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.360s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.015022] env[62235]: INFO nova.compute.claims [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.017172] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Build of instance 41adaeb9-c39f-4214-b9d1-7137aab3455f was re-scheduled: Binding failed for port 5be31e87-cd26-4ec7-a9a6-20b18809bdb1, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.018265] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.018666] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Acquiring lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.020711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Acquired lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.021690] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 618.130084] env[62235]: INFO nova.compute.manager [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] [instance: 028e94f4-6b8f-45c3-a046-ed44e575362a] Took 1.03 seconds to deallocate network for instance. [ 618.208977] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.559998] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.697520] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.741019] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.175231] env[62235]: INFO nova.scheduler.client.report [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Deleted allocations for instance 028e94f4-6b8f-45c3-a046-ed44e575362a [ 619.201584] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Releasing lock "refresh_cache-41adaeb9-c39f-4214-b9d1-7137aab3455f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.201806] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 619.201982] env[62235]: DEBUG nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.202275] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 619.239530] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.509356] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92a64d5-7fed-40f1-bef5-012982ff165b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.516949] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bf882d-1a59-46f9-a9e5-9cd99a88f4a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.162104] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e2ae20f3-b64a-476a-affb-97b4a275f6de tempest-ServerDiagnosticsTest-1074066890 tempest-ServerDiagnosticsTest-1074066890-project-member] Lock "028e94f4-6b8f-45c3-a046-ed44e575362a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.201s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.162104] env[62235]: DEBUG nova.network.neutron [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.166937] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3645ebf2-9664-4b7a-833d-66148978132b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.178230] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8226eb-93fe-438c-9c9c-2ab48a8aab09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.198387] env[62235]: DEBUG nova.compute.provider_tree [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.667069] env[62235]: INFO nova.compute.manager [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] [instance: 41adaeb9-c39f-4214-b9d1-7137aab3455f] Took 1.46 seconds to deallocate network for instance. [ 620.670521] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 620.702459] env[62235]: DEBUG nova.scheduler.client.report [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.206281] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.207063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.194s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.207531] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.212565] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.584s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.215842] env[62235]: INFO nova.compute.claims [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.702550] env[62235]: INFO nova.scheduler.client.report [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Deleted allocations for instance 41adaeb9-c39f-4214-b9d1-7137aab3455f [ 621.720960] env[62235]: DEBUG nova.compute.utils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.726192] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.726395] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 621.801687] env[62235]: DEBUG nova.policy [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19727a6117df4d298b926f707483a8e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2171f5abf1ef4e1987b9a77d132215bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.213114] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e91097de-79ef-471a-adbf-4516147afaaa tempest-ImagesOneServerNegativeTestJSON-1053686686 tempest-ImagesOneServerNegativeTestJSON-1053686686-project-member] Lock "41adaeb9-c39f-4214-b9d1-7137aab3455f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.139s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.233909] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.446320] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Successfully created port: d86a78df-3ebf-47a1-b529-583c3010167d {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.695145] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd24961e-ad72-4fd1-bf6b-a17c3b0566f7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.706643] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76825300-1dbf-4b40-af37-d38ffc2dce99 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.748662] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.756430] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fa0272-9c15-4d39-96cb-f056d5495a21 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.767864] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6131bf-e3f1-471e-af3a-de370f4fdd34 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.782954] env[62235]: DEBUG nova.compute.provider_tree [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.259240] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.286582] env[62235]: DEBUG nova.scheduler.client.report [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.292852] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.299567] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.299567] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.299567] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.299971] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.299971] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.299971] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.300089] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.300379] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.300616] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.300844] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.301071] env[62235]: DEBUG nova.virt.hardware [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.302048] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c1b2f1-fac5-4e6c-88ce-7ee7472aaa9a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.314216] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefd2e1e-0d3b-48a8-90b5-4640fcde8d19 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.794926] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.582s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.796525] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.803021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.436s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.085451] env[62235]: DEBUG nova.compute.manager [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Received event network-changed-d86a78df-3ebf-47a1-b529-583c3010167d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.085642] env[62235]: DEBUG nova.compute.manager [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Refreshing instance network info cache due to event network-changed-d86a78df-3ebf-47a1-b529-583c3010167d. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 624.086086] env[62235]: DEBUG oslo_concurrency.lockutils [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] Acquiring lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.087173] env[62235]: DEBUG oslo_concurrency.lockutils [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] Acquired lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.087173] env[62235]: DEBUG nova.network.neutron [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Refreshing network info cache for port d86a78df-3ebf-47a1-b529-583c3010167d {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 624.311191] env[62235]: DEBUG nova.compute.utils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.314586] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.315142] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 624.419811] env[62235]: DEBUG nova.policy [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f55cf69b2ce46b8851bfdc92d550b62', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e05692d31f044e859aa8abeeba5e5743', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.497938] env[62235]: ERROR nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. [ 624.497938] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.497938] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.497938] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.497938] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.497938] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.497938] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.497938] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.497938] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.497938] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 624.497938] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.497938] env[62235]: ERROR nova.compute.manager raise self.value [ 624.497938] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.497938] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.497938] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.497938] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.498337] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.498337] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.498337] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. [ 624.498337] env[62235]: ERROR nova.compute.manager [ 624.498337] env[62235]: Traceback (most recent call last): [ 624.498337] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.498337] env[62235]: listener.cb(fileno) [ 624.498337] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.498337] env[62235]: result = function(*args, **kwargs) [ 624.498337] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.498337] env[62235]: return func(*args, **kwargs) [ 624.498337] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.498337] env[62235]: raise e [ 624.498337] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.498337] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 624.498337] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.498337] env[62235]: created_port_ids = self._update_ports_for_instance( [ 624.498337] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.498337] env[62235]: with excutils.save_and_reraise_exception(): [ 624.498337] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.498337] env[62235]: self.force_reraise() [ 624.498337] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.498337] env[62235]: raise self.value [ 624.498337] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.498337] env[62235]: updated_port = self._update_port( [ 624.498337] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.498337] env[62235]: _ensure_no_port_binding_failure(port) [ 624.498337] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.498337] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.498996] env[62235]: nova.exception.PortBindingFailed: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. [ 624.498996] env[62235]: Removing descriptor: 16 [ 624.498996] env[62235]: ERROR nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Traceback (most recent call last): [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] yield resources [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self.driver.spawn(context, instance, image_meta, [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.498996] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] vm_ref = self.build_virtual_machine(instance, [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] for vif in network_info: [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return self._sync_wrapper(fn, *args, **kwargs) [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self.wait() [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self[:] = self._gt.wait() [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return self._exit_event.wait() [ 624.499342] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] result = hub.switch() [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return self.greenlet.switch() [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] result = function(*args, **kwargs) [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return func(*args, **kwargs) [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] raise e [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] nwinfo = self.network_api.allocate_for_instance( [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.499684] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] created_port_ids = self._update_ports_for_instance( [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] with excutils.save_and_reraise_exception(): [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self.force_reraise() [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] raise self.value [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] updated_port = self._update_port( [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] _ensure_no_port_binding_failure(port) [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.500066] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] raise exception.PortBindingFailed(port_id=port['id']) [ 624.500402] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] nova.exception.PortBindingFailed: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. [ 624.500402] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] [ 624.500402] env[62235]: INFO nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Terminating instance [ 624.500484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.608755] env[62235]: DEBUG nova.network.neutron [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.744804] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b95807-38cb-42fe-80c4-83df2a69a08d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.756019] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd50c293-6364-4e95-89bd-0cf27646c7ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.793206] env[62235]: DEBUG nova.network.neutron [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.795034] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa8a73e-2070-4f36-8892-b5cf57a49eaf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.803200] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc131da-ff50-4ac5-b9f7-88871bbcded6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.818584] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.821342] env[62235]: DEBUG nova.compute.provider_tree [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.041721] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Successfully created port: 40c41489-b589-4606-843c-019bd617fc61 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.300862] env[62235]: DEBUG oslo_concurrency.lockutils [req-287676e2-402c-4f58-9c41-8f813a9e5ca7 req-d0439cb4-8648-4ee5-bd51-6781a59f1fb5 service nova] Releasing lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.300862] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.300862] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.332789] env[62235]: DEBUG nova.scheduler.client.report [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.836913] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 625.839940] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.040s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.840639] env[62235]: ERROR nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Traceback (most recent call last): [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self.driver.spawn(context, instance, image_meta, [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] vm_ref = self.build_virtual_machine(instance, [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.840639] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] for vif in network_info: [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return self._sync_wrapper(fn, *args, **kwargs) [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self.wait() [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self[:] = self._gt.wait() [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return self._exit_event.wait() [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] result = hub.switch() [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.841040] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return self.greenlet.switch() [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] result = function(*args, **kwargs) [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] return func(*args, **kwargs) [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] raise e [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] nwinfo = self.network_api.allocate_for_instance( [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] created_port_ids = self._update_ports_for_instance( [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] with excutils.save_and_reraise_exception(): [ 625.841419] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] self.force_reraise() [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] raise self.value [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] updated_port = self._update_port( [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] _ensure_no_port_binding_failure(port) [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] raise exception.PortBindingFailed(port_id=port['id']) [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] nova.exception.PortBindingFailed: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. [ 625.841989] env[62235]: ERROR nova.compute.manager [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] [ 625.842735] env[62235]: DEBUG nova.compute.utils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.843819] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.847084] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.409s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.851427] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Build of instance e7684530-e42d-4a30-9c7c-28be8ff31b7b was re-scheduled: Binding failed for port 3a14258d-df5f-4a8b-8ffe-5fb1f19b024d, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.852453] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.852766] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Acquiring lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.853281] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Acquired lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.853281] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.880939] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.880939] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.880939] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.881087] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.882512] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.882512] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.883173] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.883173] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.883765] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.884234] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.884234] env[62235]: DEBUG nova.virt.hardware [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.885409] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32626b84-cfd7-4ba4-b27d-f5401fdc0b0e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.894902] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ae30e9-4098-4076-9918-96c4168f771d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.082157] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.368492] env[62235]: DEBUG nova.compute.manager [req-575e4c76-1ba0-442f-ad47-aba585e7a43f req-14991ce8-3c74-4522-a923-cd3b05638e6f service nova] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Received event network-vif-deleted-d86a78df-3ebf-47a1-b529-583c3010167d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.380430] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.491119] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.584348] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.584962] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.585036] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.585379] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8043ed1b-7734-48e2-b946-eb3440201125 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.597075] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cda29c-da66-47ca-b548-8ac47616d86f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.636994] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 063af6b5-92ce-40b0-81e0-0dd6155898fc could not be found. [ 626.636994] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.636994] env[62235]: INFO nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 626.636994] env[62235]: DEBUG oslo.service.loopingcall [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.636994] env[62235]: DEBUG nova.compute.manager [-] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.636994] env[62235]: DEBUG nova.network.neutron [-] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.665296] env[62235]: DEBUG nova.network.neutron [-] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.823106] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852033b7-3b24-4a18-b8a3-14f58b62f7e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.833035] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52243168-85b4-4989-b1d0-39d9a37fb9f7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.871974] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d64b625-eaba-4fb9-9315-b7e9970e278f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.880592] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31b82d5-ac5a-4dc8-974d-1e40d68069b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.895438] env[62235]: DEBUG nova.compute.provider_tree [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.957134] env[62235]: ERROR nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. [ 626.957134] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.957134] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.957134] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.957134] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.957134] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.957134] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.957134] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.957134] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.957134] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 626.957134] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.957134] env[62235]: ERROR nova.compute.manager raise self.value [ 626.957134] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.957134] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.957134] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.957134] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.957664] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.957664] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.957664] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. [ 626.957664] env[62235]: ERROR nova.compute.manager [ 626.957664] env[62235]: Traceback (most recent call last): [ 626.957664] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.957664] env[62235]: listener.cb(fileno) [ 626.957664] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.957664] env[62235]: result = function(*args, **kwargs) [ 626.957664] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.957664] env[62235]: return func(*args, **kwargs) [ 626.957664] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.957664] env[62235]: raise e [ 626.957664] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.957664] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 626.957664] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.957664] env[62235]: created_port_ids = self._update_ports_for_instance( [ 626.957664] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.957664] env[62235]: with excutils.save_and_reraise_exception(): [ 626.957664] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.957664] env[62235]: self.force_reraise() [ 626.957664] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.957664] env[62235]: raise self.value [ 626.957664] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.957664] env[62235]: updated_port = self._update_port( [ 626.957664] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.957664] env[62235]: _ensure_no_port_binding_failure(port) [ 626.957664] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.957664] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.958901] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. [ 626.958901] env[62235]: Removing descriptor: 22 [ 626.958901] env[62235]: ERROR nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Traceback (most recent call last): [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] yield resources [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self.driver.spawn(context, instance, image_meta, [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.958901] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] vm_ref = self.build_virtual_machine(instance, [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] for vif in network_info: [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return self._sync_wrapper(fn, *args, **kwargs) [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self.wait() [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self[:] = self._gt.wait() [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return self._exit_event.wait() [ 626.959579] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] result = hub.switch() [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return self.greenlet.switch() [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] result = function(*args, **kwargs) [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return func(*args, **kwargs) [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] raise e [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] nwinfo = self.network_api.allocate_for_instance( [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.960125] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] created_port_ids = self._update_ports_for_instance( [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] with excutils.save_and_reraise_exception(): [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self.force_reraise() [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] raise self.value [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] updated_port = self._update_port( [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] _ensure_no_port_binding_failure(port) [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.961925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] raise exception.PortBindingFailed(port_id=port['id']) [ 626.962568] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] nova.exception.PortBindingFailed: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. [ 626.962568] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] [ 626.962568] env[62235]: INFO nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Terminating instance [ 626.962723] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Acquiring lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.962755] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Acquired lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.963520] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.997048] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Releasing lock "refresh_cache-e7684530-e42d-4a30-9c7c-28be8ff31b7b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.997048] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.997048] env[62235]: DEBUG nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.997048] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.014019] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.168482] env[62235]: DEBUG nova.network.neutron [-] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.255660] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "430056ca-ccba-45c5-adf7-9c068f81f9ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.256431] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "430056ca-ccba-45c5-adf7-9c068f81f9ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.399137] env[62235]: DEBUG nova.scheduler.client.report [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.481990] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.514825] env[62235]: DEBUG nova.network.neutron [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.575547] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.671083] env[62235]: INFO nova.compute.manager [-] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Took 1.03 seconds to deallocate network for instance. [ 627.674185] env[62235]: DEBUG nova.compute.claims [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.674185] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.906337] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.906674] env[62235]: ERROR nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Traceback (most recent call last): [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self.driver.spawn(context, instance, image_meta, [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] vm_ref = self.build_virtual_machine(instance, [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.906674] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] for vif in network_info: [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return self._sync_wrapper(fn, *args, **kwargs) [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self.wait() [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self[:] = self._gt.wait() [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return self._exit_event.wait() [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] result = hub.switch() [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.907042] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return self.greenlet.switch() [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] result = function(*args, **kwargs) [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] return func(*args, **kwargs) [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] raise e [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] nwinfo = self.network_api.allocate_for_instance( [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] created_port_ids = self._update_ports_for_instance( [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] with excutils.save_and_reraise_exception(): [ 627.907391] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] self.force_reraise() [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] raise self.value [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] updated_port = self._update_port( [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] _ensure_no_port_binding_failure(port) [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] raise exception.PortBindingFailed(port_id=port['id']) [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] nova.exception.PortBindingFailed: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. [ 627.907706] env[62235]: ERROR nova.compute.manager [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] [ 627.907966] env[62235]: DEBUG nova.compute.utils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.911036] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Build of instance ec00f2ba-ed0c-47de-a9b2-bd277f96e65f was re-scheduled: Binding failed for port 09157d54-9213-410e-b1cb-8b7520390e36, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.911488] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.911726] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Acquiring lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.911891] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Acquired lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.912068] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.915701] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.525s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.018272] env[62235]: INFO nova.compute.manager [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] [instance: e7684530-e42d-4a30-9c7c-28be8ff31b7b] Took 1.02 seconds to deallocate network for instance. [ 628.082222] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Releasing lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.082668] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.082923] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 628.083297] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f2d640b-2d1a-4dc6-863a-c99f509cb0a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.099690] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0040c1bd-099b-4509-b247-3abfeeb88adb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.132020] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7adac743-926e-4c66-8fbe-02b96fbdb2b4 could not be found. [ 628.132606] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.132606] env[62235]: INFO nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 628.132737] env[62235]: DEBUG oslo.service.loopingcall [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.133091] env[62235]: DEBUG nova.compute.manager [-] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.133091] env[62235]: DEBUG nova.network.neutron [-] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.160840] env[62235]: DEBUG nova.network.neutron [-] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.233879] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.234154] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.395176] env[62235]: DEBUG nova.compute.manager [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Received event network-changed-40c41489-b589-4606-843c-019bd617fc61 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.395176] env[62235]: DEBUG nova.compute.manager [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Refreshing instance network info cache due to event network-changed-40c41489-b589-4606-843c-019bd617fc61. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.395459] env[62235]: DEBUG oslo_concurrency.lockutils [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] Acquiring lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.395459] env[62235]: DEBUG oslo_concurrency.lockutils [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] Acquired lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.395748] env[62235]: DEBUG nova.network.neutron [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Refreshing network info cache for port 40c41489-b589-4606-843c-019bd617fc61 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 628.442539] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.591020] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.664403] env[62235]: DEBUG nova.network.neutron [-] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.915135] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece9d226-2af1-4c35-8960-9ba64cada304 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.923552] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8bd2e0-3d88-4dca-875d-dd21eb03b708 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.928476] env[62235]: DEBUG nova.network.neutron [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.956911] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8047f20b-6612-4d62-9d2d-c08fe53146cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.964548] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb535f6-e926-42ba-b8bc-2e9353cb4617 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.980690] env[62235]: DEBUG nova.compute.provider_tree [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.092356] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Releasing lock "refresh_cache-ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.092592] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.092771] env[62235]: DEBUG nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.092936] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.096338] env[62235]: INFO nova.scheduler.client.report [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Deleted allocations for instance e7684530-e42d-4a30-9c7c-28be8ff31b7b [ 629.111619] env[62235]: DEBUG nova.network.neutron [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.124868] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.167790] env[62235]: INFO nova.compute.manager [-] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Took 1.03 seconds to deallocate network for instance. [ 629.170492] env[62235]: DEBUG nova.compute.claims [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.170760] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.484164] env[62235]: DEBUG nova.scheduler.client.report [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.606060] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38e84711-ba0d-4bb1-a7c5-c79ca274854d tempest-TenantUsagesTestJSON-898220646 tempest-TenantUsagesTestJSON-898220646-project-member] Lock "e7684530-e42d-4a30-9c7c-28be8ff31b7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.096s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.614149] env[62235]: DEBUG oslo_concurrency.lockutils [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] Releasing lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.614487] env[62235]: DEBUG nova.compute.manager [req-0647b3c6-483b-4312-86ba-58782971a16b req-a985099c-d234-461f-9950-371b5bde242e service nova] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Received event network-vif-deleted-40c41489-b589-4606-843c-019bd617fc61 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.631292] env[62235]: DEBUG nova.network.neutron [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.990701] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.074s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.990701] env[62235]: ERROR nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. [ 629.990701] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Traceback (most recent call last): [ 629.990701] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.990701] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self.driver.spawn(context, instance, image_meta, [ 629.990701] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.990701] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.990701] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.990701] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] vm_ref = self.build_virtual_machine(instance, [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] for vif in network_info: [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return self._sync_wrapper(fn, *args, **kwargs) [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self.wait() [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self[:] = self._gt.wait() [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return self._exit_event.wait() [ 629.991067] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] result = hub.switch() [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return self.greenlet.switch() [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] result = function(*args, **kwargs) [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] return func(*args, **kwargs) [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] raise e [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] nwinfo = self.network_api.allocate_for_instance( [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.991405] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] created_port_ids = self._update_ports_for_instance( [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] with excutils.save_and_reraise_exception(): [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] self.force_reraise() [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] raise self.value [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] updated_port = self._update_port( [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] _ensure_no_port_binding_failure(port) [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.991816] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] raise exception.PortBindingFailed(port_id=port['id']) [ 629.992174] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] nova.exception.PortBindingFailed: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. [ 629.992174] env[62235]: ERROR nova.compute.manager [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] [ 629.992174] env[62235]: DEBUG nova.compute.utils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.993827] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.468s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.995974] env[62235]: INFO nova.compute.claims [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.997842] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Build of instance a84e793b-4dbd-4b89-8fc4-94baacecd8e7 was re-scheduled: Binding failed for port 1dfea62d-9378-4e4e-99aa-46c49f11a12d, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.998334] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.998561] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.998721] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.998902] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.049273] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "7e6ebba9-58e8-4971-b9fb-8365a61d5505" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.049273] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "7e6ebba9-58e8-4971-b9fb-8365a61d5505" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.108992] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.133514] env[62235]: INFO nova.compute.manager [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] [instance: ec00f2ba-ed0c-47de-a9b2-bd277f96e65f] Took 1.04 seconds to deallocate network for instance. [ 630.532767] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.637727] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.669546] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.173017] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "refresh_cache-a84e793b-4dbd-4b89-8fc4-94baacecd8e7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.173017] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.173017] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.173017] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.185295] env[62235]: INFO nova.scheduler.client.report [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Deleted allocations for instance ec00f2ba-ed0c-47de-a9b2-bd277f96e65f [ 631.194250] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.432920] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843eb1af-260a-42f3-bae7-d99e4cc733fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.440683] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13668e1-49f0-41d3-9589-9e2ce0ac4b94 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.471056] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88eab749-5452-4205-aa85-d33cfbe0d0ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.478062] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09919b6d-e141-4fdc-9974-beb03bc6cd70 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.491106] env[62235]: DEBUG nova.compute.provider_tree [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.693664] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dbd24a20-76cb-46cc-af55-8c98712c3076 tempest-ServerAddressesTestJSON-1489097603 tempest-ServerAddressesTestJSON-1489097603-project-member] Lock "ec00f2ba-ed0c-47de-a9b2-bd277f96e65f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.395s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.697238] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.995057] env[62235]: DEBUG nova.scheduler.client.report [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.199597] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.206017] env[62235]: INFO nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: a84e793b-4dbd-4b89-8fc4-94baacecd8e7] Took 1.03 seconds to deallocate network for instance. [ 632.500039] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.500584] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.507016] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.812s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.507016] env[62235]: INFO nova.compute.claims [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.730051] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.011020] env[62235]: DEBUG nova.compute.utils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.018117] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.018117] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 633.138144] env[62235]: DEBUG nova.policy [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '523358cd3e694a9f8729c5a913dd02df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a3fab73f65e450894bb06c4b1f0eddd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.244579] env[62235]: INFO nova.scheduler.client.report [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Deleted allocations for instance a84e793b-4dbd-4b89-8fc4-94baacecd8e7 [ 633.517111] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 633.755442] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "a84e793b-4dbd-4b89-8fc4-94baacecd8e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.376s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.942505] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564bf2ea-48e4-4bcc-9a46-db1273e063e4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.955689] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc0408a-b26c-4685-b6fd-cd0409c0289f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.990620] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1433b293-80ec-4d90-840d-e7a96524add8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.998391] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bc3a92-84a4-4b1c-8bb8-e3bddaa888b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.012907] env[62235]: DEBUG nova.compute.provider_tree [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.014735] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Successfully created port: 3398a119-54e5-4d5a-b1b3-c91742140291 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.262214] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.518383] env[62235]: DEBUG nova.scheduler.client.report [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.527095] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 634.568086] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.568086] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.568086] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.568230] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.568230] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.568230] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.568230] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.568230] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.568370] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.568370] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.568370] env[62235]: DEBUG nova.virt.hardware [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.569626] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a361bde0-a91f-4d67-bc5b-d238e4356d8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.579813] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71d59ae-052a-4963-80d3-f546ffac03d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.797295] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.023080] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.023638] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.026616] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.288s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.028062] env[62235]: INFO nova.compute.claims [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.239499] env[62235]: DEBUG nova.compute.manager [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Received event network-changed-3398a119-54e5-4d5a-b1b3-c91742140291 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.239499] env[62235]: DEBUG nova.compute.manager [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Refreshing instance network info cache due to event network-changed-3398a119-54e5-4d5a-b1b3-c91742140291. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 635.239499] env[62235]: DEBUG oslo_concurrency.lockutils [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] Acquiring lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.239499] env[62235]: DEBUG oslo_concurrency.lockutils [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] Acquired lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.239499] env[62235]: DEBUG nova.network.neutron [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Refreshing network info cache for port 3398a119-54e5-4d5a-b1b3-c91742140291 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 635.499315] env[62235]: ERROR nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. [ 635.499315] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.499315] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.499315] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.499315] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.499315] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.499315] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.499315] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.499315] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.499315] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 635.499315] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.499315] env[62235]: ERROR nova.compute.manager raise self.value [ 635.499315] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.499315] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.499315] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.499315] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.499741] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.499741] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.499741] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. [ 635.499741] env[62235]: ERROR nova.compute.manager [ 635.499741] env[62235]: Traceback (most recent call last): [ 635.499741] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.499741] env[62235]: listener.cb(fileno) [ 635.499741] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.499741] env[62235]: result = function(*args, **kwargs) [ 635.499741] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.499741] env[62235]: return func(*args, **kwargs) [ 635.499741] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.499741] env[62235]: raise e [ 635.499741] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.499741] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 635.499741] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.499741] env[62235]: created_port_ids = self._update_ports_for_instance( [ 635.499741] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.499741] env[62235]: with excutils.save_and_reraise_exception(): [ 635.499741] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.499741] env[62235]: self.force_reraise() [ 635.499741] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.499741] env[62235]: raise self.value [ 635.499741] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.499741] env[62235]: updated_port = self._update_port( [ 635.499741] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.499741] env[62235]: _ensure_no_port_binding_failure(port) [ 635.499741] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.499741] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.500418] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. [ 635.500418] env[62235]: Removing descriptor: 16 [ 635.500542] env[62235]: ERROR nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Traceback (most recent call last): [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] yield resources [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self.driver.spawn(context, instance, image_meta, [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] vm_ref = self.build_virtual_machine(instance, [ 635.500542] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] for vif in network_info: [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return self._sync_wrapper(fn, *args, **kwargs) [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self.wait() [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self[:] = self._gt.wait() [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return self._exit_event.wait() [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.500799] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] result = hub.switch() [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return self.greenlet.switch() [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] result = function(*args, **kwargs) [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return func(*args, **kwargs) [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] raise e [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] nwinfo = self.network_api.allocate_for_instance( [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] created_port_ids = self._update_ports_for_instance( [ 635.501132] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] with excutils.save_and_reraise_exception(): [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self.force_reraise() [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] raise self.value [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] updated_port = self._update_port( [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] _ensure_no_port_binding_failure(port) [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] raise exception.PortBindingFailed(port_id=port['id']) [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] nova.exception.PortBindingFailed: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. [ 635.501438] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] [ 635.501789] env[62235]: INFO nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Terminating instance [ 635.502997] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Acquiring lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.535763] env[62235]: DEBUG nova.compute.utils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.541672] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 635.541934] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 635.623667] env[62235]: DEBUG nova.policy [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '242e31f81b7944c3900af425422d1157', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd90b53cbac954198863c36ef627925b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.769147] env[62235]: DEBUG nova.network.neutron [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.964759] env[62235]: DEBUG nova.network.neutron [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.046578] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.071414] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Successfully created port: f1d4d61e-8bb9-4979-abef-995e1da9c4a6 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.380280] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.380520] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.468556] env[62235]: DEBUG oslo_concurrency.lockutils [req-8e2f54de-6846-460b-ae56-84114d45506c req-3c4177ab-2cb2-401b-87a3-cd416e29e2a3 service nova] Releasing lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.469039] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Acquired lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.469234] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.486760] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae871f0-6c5e-49b3-998d-23d589889070 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.495286] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8d5aa1-7226-462a-80e7-c6e58d2e2529 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.529595] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1053981b-75d2-47b0-b729-49ceec88ad11 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.537204] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10dbb05d-ad20-423f-b1c0-708280bbe526 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.557936] env[62235]: DEBUG nova.compute.provider_tree [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.994161] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.060279] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.066300] env[62235]: DEBUG nova.scheduler.client.report [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.097346] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.097603] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.097753] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.097927] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.100182] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.100506] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.100851] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.101156] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.101872] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.101872] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.101872] env[62235]: DEBUG nova.virt.hardware [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.102967] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a841f05c-f082-4a1b-b83a-70e0eec02cc7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.112225] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2097ab80-0f39-4f84-b781-e6e35b7ef7cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.163661] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.235342] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Acquiring lock "6671163c-c829-4046-97e5-e85fe5e223dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.235342] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Lock "6671163c-c829-4046-97e5-e85fe5e223dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.265625] env[62235]: DEBUG nova.compute.manager [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Received event network-vif-deleted-3398a119-54e5-4d5a-b1b3-c91742140291 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.265838] env[62235]: DEBUG nova.compute.manager [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Received event network-changed-f1d4d61e-8bb9-4979-abef-995e1da9c4a6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.266011] env[62235]: DEBUG nova.compute.manager [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Refreshing instance network info cache due to event network-changed-f1d4d61e-8bb9-4979-abef-995e1da9c4a6. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.266255] env[62235]: DEBUG oslo_concurrency.lockutils [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] Acquiring lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.266361] env[62235]: DEBUG oslo_concurrency.lockutils [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] Acquired lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.266515] env[62235]: DEBUG nova.network.neutron [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Refreshing network info cache for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.324899] env[62235]: ERROR nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. [ 637.324899] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.324899] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.324899] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.324899] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.324899] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.324899] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.324899] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.324899] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.324899] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 637.324899] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.324899] env[62235]: ERROR nova.compute.manager raise self.value [ 637.324899] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.324899] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.324899] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.324899] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.325312] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.325312] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.325312] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. [ 637.325312] env[62235]: ERROR nova.compute.manager [ 637.325312] env[62235]: Traceback (most recent call last): [ 637.325312] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.325312] env[62235]: listener.cb(fileno) [ 637.325312] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.325312] env[62235]: result = function(*args, **kwargs) [ 637.325312] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.325312] env[62235]: return func(*args, **kwargs) [ 637.325312] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.325312] env[62235]: raise e [ 637.325312] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.325312] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 637.325312] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.325312] env[62235]: created_port_ids = self._update_ports_for_instance( [ 637.325312] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.325312] env[62235]: with excutils.save_and_reraise_exception(): [ 637.325312] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.325312] env[62235]: self.force_reraise() [ 637.325312] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.325312] env[62235]: raise self.value [ 637.325312] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.325312] env[62235]: updated_port = self._update_port( [ 637.325312] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.325312] env[62235]: _ensure_no_port_binding_failure(port) [ 637.325312] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.325312] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.326021] env[62235]: nova.exception.PortBindingFailed: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. [ 637.326021] env[62235]: Removing descriptor: 16 [ 637.326021] env[62235]: ERROR nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Traceback (most recent call last): [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] yield resources [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self.driver.spawn(context, instance, image_meta, [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.326021] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] vm_ref = self.build_virtual_machine(instance, [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] for vif in network_info: [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return self._sync_wrapper(fn, *args, **kwargs) [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self.wait() [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self[:] = self._gt.wait() [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return self._exit_event.wait() [ 637.326303] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] result = hub.switch() [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return self.greenlet.switch() [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] result = function(*args, **kwargs) [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return func(*args, **kwargs) [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] raise e [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] nwinfo = self.network_api.allocate_for_instance( [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.326608] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] created_port_ids = self._update_ports_for_instance( [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] with excutils.save_and_reraise_exception(): [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self.force_reraise() [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] raise self.value [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] updated_port = self._update_port( [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] _ensure_no_port_binding_failure(port) [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.327090] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] raise exception.PortBindingFailed(port_id=port['id']) [ 637.327374] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] nova.exception.PortBindingFailed: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. [ 637.327374] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] [ 637.327374] env[62235]: INFO nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Terminating instance [ 637.328061] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.575839] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.576304] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.580572] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.373s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.580865] env[62235]: INFO nova.compute.claims [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.666584] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Releasing lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.667094] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.667334] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.667619] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4dc01839-b8f3-4a6e-b829-ca297e76b06e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.676817] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415103e4-04b0-4d09-b8c6-9e55eb37a795 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.699876] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 74d069ed-c262-4b91-bcdd-ea3b4121adc4 could not be found. [ 637.699876] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.699876] env[62235]: INFO nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 637.699876] env[62235]: DEBUG oslo.service.loopingcall [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.700171] env[62235]: DEBUG nova.compute.manager [-] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.700171] env[62235]: DEBUG nova.network.neutron [-] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.714545] env[62235]: DEBUG nova.network.neutron [-] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.793391] env[62235]: DEBUG nova.network.neutron [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.850460] env[62235]: DEBUG nova.network.neutron [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.085284] env[62235]: DEBUG nova.compute.utils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.087058] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 638.087235] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 638.140250] env[62235]: DEBUG nova.policy [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3eaa842ad5bb4567b29ae2d72f89b115', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8edfe84cd217446dbe3dbedce3943030', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 638.216746] env[62235]: DEBUG nova.network.neutron [-] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.353582] env[62235]: DEBUG oslo_concurrency.lockutils [req-f538ad3a-5a49-4fc0-ad64-2f446717379e req-2912ef91-02e2-435d-a151-70018572b1a0 service nova] Releasing lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.354067] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquired lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.354266] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.430396] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Successfully created port: e230395f-db41-4e33-aa62-c1593806f605 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.595018] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.719585] env[62235]: INFO nova.compute.manager [-] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Took 1.02 seconds to deallocate network for instance. [ 638.725674] env[62235]: DEBUG nova.compute.claims [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.726021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.875559] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.956988] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.995749] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c50ebe3-fd2e-4683-8dc1-27795c4cfd82 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.003696] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bd0110-74e3-434e-a6bf-930c22e4057c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.034967] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4883ae69-b923-4cf3-995f-f5f7b75fd834 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.042809] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0633c7-79f9-472e-9edd-eb9ac7ebb979 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.055195] env[62235]: DEBUG nova.compute.provider_tree [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.287672] env[62235]: DEBUG nova.compute.manager [req-3bb38d3e-6ec3-4981-b842-8e5f7f5a3e05 req-5802ec6e-485f-42c7-84e3-f221640865e1 service nova] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Received event network-vif-deleted-f1d4d61e-8bb9-4979-abef-995e1da9c4a6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.462310] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Releasing lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.462722] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.463487] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.463487] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-298341ff-36ce-48a9-8517-d25b91519323 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.471990] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda7f172-cce5-41d2-9b24-d9d1a6f452ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.493650] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4 could not be found. [ 639.493858] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.494057] env[62235]: INFO nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 639.494304] env[62235]: DEBUG oslo.service.loopingcall [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.494505] env[62235]: DEBUG nova.compute.manager [-] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.494597] env[62235]: DEBUG nova.network.neutron [-] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.514970] env[62235]: DEBUG nova.network.neutron [-] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.559452] env[62235]: DEBUG nova.scheduler.client.report [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.582777] env[62235]: ERROR nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. [ 639.582777] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 639.582777] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.582777] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 639.582777] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.582777] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 639.582777] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.582777] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 639.582777] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.582777] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 639.582777] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.582777] env[62235]: ERROR nova.compute.manager raise self.value [ 639.582777] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.582777] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 639.582777] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.582777] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 639.583354] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.583354] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 639.583354] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. [ 639.583354] env[62235]: ERROR nova.compute.manager [ 639.583354] env[62235]: Traceback (most recent call last): [ 639.583354] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 639.583354] env[62235]: listener.cb(fileno) [ 639.583354] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.583354] env[62235]: result = function(*args, **kwargs) [ 639.583354] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.583354] env[62235]: return func(*args, **kwargs) [ 639.583354] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.583354] env[62235]: raise e [ 639.583354] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.583354] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 639.583354] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.583354] env[62235]: created_port_ids = self._update_ports_for_instance( [ 639.583354] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.583354] env[62235]: with excutils.save_and_reraise_exception(): [ 639.583354] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.583354] env[62235]: self.force_reraise() [ 639.583354] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.583354] env[62235]: raise self.value [ 639.583354] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.583354] env[62235]: updated_port = self._update_port( [ 639.583354] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.583354] env[62235]: _ensure_no_port_binding_failure(port) [ 639.583354] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.583354] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 639.584152] env[62235]: nova.exception.PortBindingFailed: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. [ 639.584152] env[62235]: Removing descriptor: 16 [ 639.603769] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.628187] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.628432] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.628591] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.628790] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.628960] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.629131] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.629339] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.629557] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.629661] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.629821] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.629990] env[62235]: DEBUG nova.virt.hardware [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.630906] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842b2d6b-a516-4b0b-9a5c-63b942c7653e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.641308] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3101ea-556b-40b2-8798-373c5fad8616 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.656734] env[62235]: ERROR nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Traceback (most recent call last): [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] yield resources [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self.driver.spawn(context, instance, image_meta, [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] vm_ref = self.build_virtual_machine(instance, [ 639.656734] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] for vif in network_info: [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] return self._sync_wrapper(fn, *args, **kwargs) [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self.wait() [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self[:] = self._gt.wait() [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] return self._exit_event.wait() [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 639.657240] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] current.throw(*self._exc) [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] result = function(*args, **kwargs) [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] return func(*args, **kwargs) [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] raise e [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] nwinfo = self.network_api.allocate_for_instance( [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] created_port_ids = self._update_ports_for_instance( [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] with excutils.save_and_reraise_exception(): [ 639.657553] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self.force_reraise() [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] raise self.value [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] updated_port = self._update_port( [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] _ensure_no_port_binding_failure(port) [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] raise exception.PortBindingFailed(port_id=port['id']) [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] nova.exception.PortBindingFailed: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. [ 639.657847] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] [ 639.657847] env[62235]: INFO nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Terminating instance [ 639.658680] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.658876] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquired lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.659068] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.017715] env[62235]: DEBUG nova.network.neutron [-] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.056075] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 640.056502] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 640.064965] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.065431] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.068066] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.776s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.069491] env[62235]: INFO nova.compute.claims [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.178504] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.254188] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.520843] env[62235]: INFO nova.compute.manager [-] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Took 1.03 seconds to deallocate network for instance. [ 640.523233] env[62235]: DEBUG nova.compute.claims [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.523415] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.562233] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 640.562412] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 640.575455] env[62235]: DEBUG nova.compute.utils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.578832] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.578995] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 640.618647] env[62235]: DEBUG nova.policy [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3eaa842ad5bb4567b29ae2d72f89b115', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8edfe84cd217446dbe3dbedce3943030', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.756706] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Releasing lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.757159] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 640.757348] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.757653] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32609ce8-2a01-4d13-a9be-1d8b51cf9f69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.769943] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8ad699-e96a-460a-a693-02fb15e435d5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.792525] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91e05f87-96a4-4d52-9016-841a3a034930 could not be found. [ 640.792732] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 640.792913] env[62235]: INFO nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Took 0.04 seconds to destroy the instance on the hypervisor. [ 640.793185] env[62235]: DEBUG oslo.service.loopingcall [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.793413] env[62235]: DEBUG nova.compute.manager [-] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.793516] env[62235]: DEBUG nova.network.neutron [-] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.828572] env[62235]: DEBUG nova.network.neutron [-] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.939950] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Successfully created port: b8a5a8db-3cae-42a6-a67b-2c42b40ae081 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.064937] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Didn't find any instances for network info cache update. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 641.065458] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.065833] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.066126] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.066395] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.066635] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.066892] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.067264] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 641.067432] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.079931] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.331316] env[62235]: DEBUG nova.network.neutron [-] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.344152] env[62235]: DEBUG nova.compute.manager [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Received event network-changed-e230395f-db41-4e33-aa62-c1593806f605 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.344362] env[62235]: DEBUG nova.compute.manager [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Refreshing instance network info cache due to event network-changed-e230395f-db41-4e33-aa62-c1593806f605. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.344568] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] Acquiring lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.344706] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] Acquired lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.344861] env[62235]: DEBUG nova.network.neutron [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Refreshing network info cache for port e230395f-db41-4e33-aa62-c1593806f605 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.571636] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.602991] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985c1813-0c9d-4923-9824-aa6a95d5c727 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.612591] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61701f3f-f50e-4c11-9b1f-a13f933863f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.643773] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabd3322-e271-41f5-9a6b-d8669616b34b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.651419] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6654d59d-1215-42eb-b58e-12cd3ad5f270 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.665565] env[62235]: DEBUG nova.compute.provider_tree [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.847105] env[62235]: INFO nova.compute.manager [-] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Took 1.05 seconds to deallocate network for instance. [ 641.852362] env[62235]: DEBUG nova.compute.claims [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.852557] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.866834] env[62235]: DEBUG nova.network.neutron [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.951961] env[62235]: ERROR nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. [ 641.951961] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.951961] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.951961] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.951961] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.951961] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.951961] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.951961] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.951961] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.951961] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 641.951961] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.951961] env[62235]: ERROR nova.compute.manager raise self.value [ 641.951961] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.951961] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.951961] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.951961] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.952393] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.952393] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.952393] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. [ 641.952393] env[62235]: ERROR nova.compute.manager [ 641.952393] env[62235]: Traceback (most recent call last): [ 641.952393] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.952393] env[62235]: listener.cb(fileno) [ 641.952393] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.952393] env[62235]: result = function(*args, **kwargs) [ 641.952393] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.952393] env[62235]: return func(*args, **kwargs) [ 641.952393] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.952393] env[62235]: raise e [ 641.952393] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.952393] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 641.952393] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.952393] env[62235]: created_port_ids = self._update_ports_for_instance( [ 641.952393] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.952393] env[62235]: with excutils.save_and_reraise_exception(): [ 641.952393] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.952393] env[62235]: self.force_reraise() [ 641.952393] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.952393] env[62235]: raise self.value [ 641.952393] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.952393] env[62235]: updated_port = self._update_port( [ 641.952393] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.952393] env[62235]: _ensure_no_port_binding_failure(port) [ 641.952393] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.952393] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.953119] env[62235]: nova.exception.PortBindingFailed: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. [ 641.953119] env[62235]: Removing descriptor: 16 [ 641.999082] env[62235]: DEBUG nova.network.neutron [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.097180] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.123830] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.124107] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.124271] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.124454] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.124596] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.124741] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.125148] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.125148] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.125261] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.125417] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.125585] env[62235]: DEBUG nova.virt.hardware [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.126461] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a72356-5295-40f7-a978-d9bd5361ef19 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.134941] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd9470e-d490-438f-82a9-2b2c772ec510 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.148607] env[62235]: ERROR nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Traceback (most recent call last): [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] yield resources [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self.driver.spawn(context, instance, image_meta, [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] vm_ref = self.build_virtual_machine(instance, [ 642.148607] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] for vif in network_info: [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] return self._sync_wrapper(fn, *args, **kwargs) [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self.wait() [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self[:] = self._gt.wait() [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] return self._exit_event.wait() [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.149064] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] current.throw(*self._exc) [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] result = function(*args, **kwargs) [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] return func(*args, **kwargs) [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] raise e [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] nwinfo = self.network_api.allocate_for_instance( [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] created_port_ids = self._update_ports_for_instance( [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] with excutils.save_and_reraise_exception(): [ 642.149496] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self.force_reraise() [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] raise self.value [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] updated_port = self._update_port( [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] _ensure_no_port_binding_failure(port) [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] raise exception.PortBindingFailed(port_id=port['id']) [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] nova.exception.PortBindingFailed: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. [ 642.149850] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] [ 642.149850] env[62235]: INFO nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Terminating instance [ 642.151612] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.151767] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquired lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.151927] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.168025] env[62235]: DEBUG nova.scheduler.client.report [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.501649] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] Releasing lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.503040] env[62235]: DEBUG nova.compute.manager [req-ba5ca389-c442-4805-b8c3-53acafe67866 req-e3f2e764-bba7-4398-b17f-fa03d20092a7 service nova] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Received event network-vif-deleted-e230395f-db41-4e33-aa62-c1593806f605 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 642.668404] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.673180] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.673676] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.676205] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.002s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.740344] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.181749] env[62235]: DEBUG nova.compute.utils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.185712] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.185875] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.224993] env[62235]: DEBUG nova.policy [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efee98a77766497ca4ac12c818058a44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6def7fdd3ad4de89aacfba958f0b23a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.242500] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Releasing lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.243263] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.243263] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.243395] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af760f35-d9b6-478f-9222-9a4a63f6e85d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.256847] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f1c257-6744-46cb-aaef-1b4c0ca8ba91 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.292455] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 279097d2-348d-4f58-88ec-8e2e3143f754 could not be found. [ 643.292809] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.293076] env[62235]: INFO nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Took 0.05 seconds to destroy the instance on the hypervisor. [ 643.293416] env[62235]: DEBUG oslo.service.loopingcall [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.296842] env[62235]: DEBUG nova.compute.manager [-] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.296968] env[62235]: DEBUG nova.network.neutron [-] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.317226] env[62235]: DEBUG nova.network.neutron [-] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.378544] env[62235]: DEBUG nova.compute.manager [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Received event network-changed-b8a5a8db-3cae-42a6-a67b-2c42b40ae081 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.378739] env[62235]: DEBUG nova.compute.manager [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Refreshing instance network info cache due to event network-changed-b8a5a8db-3cae-42a6-a67b-2c42b40ae081. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 643.378957] env[62235]: DEBUG oslo_concurrency.lockutils [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] Acquiring lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.379133] env[62235]: DEBUG oslo_concurrency.lockutils [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] Acquired lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.379479] env[62235]: DEBUG nova.network.neutron [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Refreshing network info cache for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 643.566304] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Successfully created port: 58b8b07e-114d-40bb-a937-07d349e00319 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.601611] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e381429-64fb-4230-849d-18039c9c39a9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.612036] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a8b538-ad44-4b73-a674-45b1130fc551 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.639587] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff213ea8-b788-41fa-adab-5eb14c88ed75 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.646961] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3f679b-e97a-453a-857c-f7321f5a91fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.660730] env[62235]: DEBUG nova.compute.provider_tree [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.688158] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 643.819478] env[62235]: DEBUG nova.network.neutron [-] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.898998] env[62235]: DEBUG nova.network.neutron [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.020436] env[62235]: DEBUG nova.network.neutron [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.163574] env[62235]: DEBUG nova.scheduler.client.report [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.322025] env[62235]: INFO nova.compute.manager [-] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Took 1.02 seconds to deallocate network for instance. [ 644.324846] env[62235]: DEBUG nova.compute.claims [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.326029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.524563] env[62235]: DEBUG oslo_concurrency.lockutils [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] Releasing lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.524647] env[62235]: DEBUG nova.compute.manager [req-f762f841-9d3a-46ce-a655-897b0decb611 req-2eb66a2f-d7e9-4ba3-9f9c-624abbc7f462 service nova] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Received event network-vif-deleted-b8a5a8db-3cae-42a6-a67b-2c42b40ae081 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.542855] env[62235]: ERROR nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. [ 644.542855] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.542855] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.542855] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.542855] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.542855] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.542855] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.542855] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.542855] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.542855] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 644.542855] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.542855] env[62235]: ERROR nova.compute.manager raise self.value [ 644.542855] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.542855] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.542855] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.542855] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.543497] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.543497] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.543497] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. [ 644.543497] env[62235]: ERROR nova.compute.manager [ 644.543497] env[62235]: Traceback (most recent call last): [ 644.543497] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.543497] env[62235]: listener.cb(fileno) [ 644.543497] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.543497] env[62235]: result = function(*args, **kwargs) [ 644.543497] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.543497] env[62235]: return func(*args, **kwargs) [ 644.543497] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.543497] env[62235]: raise e [ 644.543497] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.543497] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 644.543497] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.543497] env[62235]: created_port_ids = self._update_ports_for_instance( [ 644.543497] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.543497] env[62235]: with excutils.save_and_reraise_exception(): [ 644.543497] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.543497] env[62235]: self.force_reraise() [ 644.543497] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.543497] env[62235]: raise self.value [ 644.543497] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.543497] env[62235]: updated_port = self._update_port( [ 644.543497] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.543497] env[62235]: _ensure_no_port_binding_failure(port) [ 644.543497] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.543497] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.544157] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. [ 644.544157] env[62235]: Removing descriptor: 22 [ 644.669561] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.670678] env[62235]: ERROR nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Traceback (most recent call last): [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self.driver.spawn(context, instance, image_meta, [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] vm_ref = self.build_virtual_machine(instance, [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.670678] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] for vif in network_info: [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return self._sync_wrapper(fn, *args, **kwargs) [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self.wait() [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self[:] = self._gt.wait() [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return self._exit_event.wait() [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] result = hub.switch() [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.671132] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return self.greenlet.switch() [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] result = function(*args, **kwargs) [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] return func(*args, **kwargs) [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] raise e [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] nwinfo = self.network_api.allocate_for_instance( [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] created_port_ids = self._update_ports_for_instance( [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] with excutils.save_and_reraise_exception(): [ 644.671491] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] self.force_reraise() [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] raise self.value [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] updated_port = self._update_port( [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] _ensure_no_port_binding_failure(port) [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] raise exception.PortBindingFailed(port_id=port['id']) [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] nova.exception.PortBindingFailed: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. [ 644.671815] env[62235]: ERROR nova.compute.manager [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] [ 644.672109] env[62235]: DEBUG nova.compute.utils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.672264] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.502s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.675550] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Build of instance 063af6b5-92ce-40b0-81e0-0dd6155898fc was re-scheduled: Binding failed for port d86a78df-3ebf-47a1-b529-583c3010167d, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 644.676915] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 644.676915] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.676915] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.676915] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 644.696528] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 644.724422] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 644.724695] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 644.724849] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.725039] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 644.725189] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.725414] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 644.725522] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 644.725678] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 644.725841] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 644.726007] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 644.726192] env[62235]: DEBUG nova.virt.hardware [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.727053] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a8cd3c-58c7-4b21-914f-9ae29420025f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.735333] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e91373e-ba44-4cd3-99a9-1f1eb7d22287 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.750092] env[62235]: ERROR nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Traceback (most recent call last): [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] yield resources [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self.driver.spawn(context, instance, image_meta, [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] vm_ref = self.build_virtual_machine(instance, [ 644.750092] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] for vif in network_info: [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] return self._sync_wrapper(fn, *args, **kwargs) [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self.wait() [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self[:] = self._gt.wait() [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] return self._exit_event.wait() [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 644.750507] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] current.throw(*self._exc) [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] result = function(*args, **kwargs) [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] return func(*args, **kwargs) [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] raise e [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] nwinfo = self.network_api.allocate_for_instance( [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] created_port_ids = self._update_ports_for_instance( [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] with excutils.save_and_reraise_exception(): [ 644.750879] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self.force_reraise() [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] raise self.value [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] updated_port = self._update_port( [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] _ensure_no_port_binding_failure(port) [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] raise exception.PortBindingFailed(port_id=port['id']) [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] nova.exception.PortBindingFailed: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. [ 644.751274] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] [ 644.751274] env[62235]: INFO nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Terminating instance [ 644.752357] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Acquiring lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.752513] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Acquired lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.752675] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.195716] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.285102] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.340785] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.405221] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.421350] env[62235]: DEBUG nova.compute.manager [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Received event network-changed-58b8b07e-114d-40bb-a937-07d349e00319 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 645.421536] env[62235]: DEBUG nova.compute.manager [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Refreshing instance network info cache due to event network-changed-58b8b07e-114d-40bb-a937-07d349e00319. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 645.421648] env[62235]: DEBUG oslo_concurrency.lockutils [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] Acquiring lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.525191] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caedc096-da9a-4db3-bddd-942b47f155d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.532529] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6177a4-e580-4860-91ed-39371f0dd6f3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.563368] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16853d6-8591-42fb-a4fe-ca4cff50b0de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.570759] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7195d185-dc2f-401c-a823-379a5ac41a34 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.585348] env[62235]: DEBUG nova.compute.provider_tree [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.846099] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "refresh_cache-063af6b5-92ce-40b0-81e0-0dd6155898fc" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.846388] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 645.846538] env[62235]: DEBUG nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.846763] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 645.861698] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.908142] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Releasing lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.908593] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.908789] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 645.909115] env[62235]: DEBUG oslo_concurrency.lockutils [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] Acquired lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.909289] env[62235]: DEBUG nova.network.neutron [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Refreshing network info cache for port 58b8b07e-114d-40bb-a937-07d349e00319 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 645.910278] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f687f5e-4f02-4efa-9b9f-da5937da6390 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.919675] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c8eb66-943b-4d43-806e-3d78136f2ddf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.940994] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1d1a57fc-6bc7-47a1-918e-8d23ac797d54 could not be found. [ 645.941231] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.941410] env[62235]: INFO nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Took 0.03 seconds to destroy the instance on the hypervisor. [ 645.941649] env[62235]: DEBUG oslo.service.loopingcall [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.941870] env[62235]: DEBUG nova.compute.manager [-] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.941979] env[62235]: DEBUG nova.network.neutron [-] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 645.957113] env[62235]: DEBUG nova.network.neutron [-] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.088352] env[62235]: DEBUG nova.scheduler.client.report [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.364304] env[62235]: DEBUG nova.network.neutron [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.430194] env[62235]: DEBUG nova.network.neutron [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.459982] env[62235]: DEBUG nova.network.neutron [-] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.530836] env[62235]: DEBUG nova.network.neutron [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.593277] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.593925] env[62235]: ERROR nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Traceback (most recent call last): [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self.driver.spawn(context, instance, image_meta, [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] vm_ref = self.build_virtual_machine(instance, [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.593925] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] for vif in network_info: [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return self._sync_wrapper(fn, *args, **kwargs) [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self.wait() [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self[:] = self._gt.wait() [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return self._exit_event.wait() [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] result = hub.switch() [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.594272] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return self.greenlet.switch() [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] result = function(*args, **kwargs) [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] return func(*args, **kwargs) [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] raise e [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] nwinfo = self.network_api.allocate_for_instance( [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] created_port_ids = self._update_ports_for_instance( [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] with excutils.save_and_reraise_exception(): [ 646.594682] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] self.force_reraise() [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] raise self.value [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] updated_port = self._update_port( [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] _ensure_no_port_binding_failure(port) [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] raise exception.PortBindingFailed(port_id=port['id']) [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] nova.exception.PortBindingFailed: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. [ 646.595053] env[62235]: ERROR nova.compute.manager [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] [ 646.595365] env[62235]: DEBUG nova.compute.utils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.595980] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.958s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.597299] env[62235]: INFO nova.compute.claims [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.599846] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Build of instance 7adac743-926e-4c66-8fbe-02b96fbdb2b4 was re-scheduled: Binding failed for port 40c41489-b589-4606-843c-019bd617fc61, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 646.600285] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 646.600505] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Acquiring lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.601434] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Acquired lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.601434] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 646.867191] env[62235]: INFO nova.compute.manager [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 063af6b5-92ce-40b0-81e0-0dd6155898fc] Took 1.02 seconds to deallocate network for instance. [ 646.962881] env[62235]: INFO nova.compute.manager [-] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Took 1.02 seconds to deallocate network for instance. [ 646.965582] env[62235]: DEBUG nova.compute.claims [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 646.965758] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.033978] env[62235]: DEBUG oslo_concurrency.lockutils [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] Releasing lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.034281] env[62235]: DEBUG nova.compute.manager [req-2f7c844d-fdba-4203-b22a-f9c6ad21ef5f req-2dfc8909-051a-4f1c-9f3f-593267c667f4 service nova] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Received event network-vif-deleted-58b8b07e-114d-40bb-a937-07d349e00319 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.120811] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.368615] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.871601] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Releasing lock "refresh_cache-7adac743-926e-4c66-8fbe-02b96fbdb2b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.872068] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 647.872068] env[62235]: DEBUG nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.872264] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 647.891728] env[62235]: INFO nova.scheduler.client.report [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Deleted allocations for instance 063af6b5-92ce-40b0-81e0-0dd6155898fc [ 647.897827] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.013733] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e693e22e-a81e-4a65-b87b-b2290352a134 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.021224] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fafb8f-6144-4512-b508-b9269f593d7e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.050620] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab11ba11-0ad1-480c-a554-eb3cc4863c07 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.057875] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa128228-f9dd-4a59-b60b-db819dde62f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.071248] env[62235]: DEBUG nova.compute.provider_tree [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.400609] env[62235]: DEBUG nova.network.neutron [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.402711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ec28ea76-64ce-46fd-9cae-7e77b49c657f tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "063af6b5-92ce-40b0-81e0-0dd6155898fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.972s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.574058] env[62235]: DEBUG nova.scheduler.client.report [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.903278] env[62235]: INFO nova.compute.manager [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] [instance: 7adac743-926e-4c66-8fbe-02b96fbdb2b4] Took 1.03 seconds to deallocate network for instance. [ 648.910023] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 649.078995] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.079538] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.082381] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.353s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.083843] env[62235]: INFO nova.compute.claims [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.435137] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.588672] env[62235]: DEBUG nova.compute.utils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.592921] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.593082] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 649.636806] env[62235]: DEBUG nova.policy [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '300b545041b24da3a3ca526032be95a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd0207bb65754d90ab4be6255c98957d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 649.935271] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Successfully created port: 8abc4405-49b4-4cc9-adfc-ff24ba7c3676 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.938324] env[62235]: INFO nova.scheduler.client.report [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Deleted allocations for instance 7adac743-926e-4c66-8fbe-02b96fbdb2b4 [ 650.097305] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.107118] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "3e7a81ae-2163-4562-b483-dc639c602fa0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.107118] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "3e7a81ae-2163-4562-b483-dc639c602fa0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.134374] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "dd7268d1-1da5-4e5f-9d45-816df147b564" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.134614] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.446299] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02a56a52-da9d-4955-8d0c-0b57480a1165 tempest-ServerActionsTestOtherB-670330660 tempest-ServerActionsTestOtherB-670330660-project-member] Lock "7adac743-926e-4c66-8fbe-02b96fbdb2b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.846s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.467048] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95217b9-9839-4e97-a234-a4397ed6a83c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.474650] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998a6595-1918-476b-9629-dcf3015b39df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.510088] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c02921-7cb7-487e-9a40-07f814b45211 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.519098] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bf9c70-5077-4ead-a56f-0455973e5d1c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.533781] env[62235]: DEBUG nova.compute.provider_tree [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.801876] env[62235]: DEBUG nova.compute.manager [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Received event network-changed-8abc4405-49b4-4cc9-adfc-ff24ba7c3676 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.802107] env[62235]: DEBUG nova.compute.manager [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Refreshing instance network info cache due to event network-changed-8abc4405-49b4-4cc9-adfc-ff24ba7c3676. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 650.802332] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] Acquiring lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.802489] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] Acquired lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.802630] env[62235]: DEBUG nova.network.neutron [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Refreshing network info cache for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.951035] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 651.036753] env[62235]: DEBUG nova.scheduler.client.report [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.084953] env[62235]: ERROR nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. [ 651.084953] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.084953] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.084953] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.084953] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.084953] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.084953] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.084953] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.084953] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.084953] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 651.084953] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.084953] env[62235]: ERROR nova.compute.manager raise self.value [ 651.084953] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.084953] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.084953] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.084953] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.085426] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.085426] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.085426] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. [ 651.085426] env[62235]: ERROR nova.compute.manager [ 651.085426] env[62235]: Traceback (most recent call last): [ 651.085426] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.085426] env[62235]: listener.cb(fileno) [ 651.085426] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.085426] env[62235]: result = function(*args, **kwargs) [ 651.085426] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.085426] env[62235]: return func(*args, **kwargs) [ 651.085426] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.085426] env[62235]: raise e [ 651.085426] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.085426] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 651.085426] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.085426] env[62235]: created_port_ids = self._update_ports_for_instance( [ 651.085426] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.085426] env[62235]: with excutils.save_and_reraise_exception(): [ 651.085426] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.085426] env[62235]: self.force_reraise() [ 651.085426] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.085426] env[62235]: raise self.value [ 651.085426] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.085426] env[62235]: updated_port = self._update_port( [ 651.085426] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.085426] env[62235]: _ensure_no_port_binding_failure(port) [ 651.085426] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.085426] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.086252] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. [ 651.086252] env[62235]: Removing descriptor: 16 [ 651.107536] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.144255] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.144505] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.144663] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.144847] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.144994] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.145429] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.145652] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.145810] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.145979] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.146159] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.146334] env[62235]: DEBUG nova.virt.hardware [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.147204] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8cb82e-6a3c-4228-af00-68ffc6085ef1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.155679] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37628850-29a2-4eed-ab08-38ea671ed545 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.171170] env[62235]: ERROR nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Traceback (most recent call last): [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] yield resources [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self.driver.spawn(context, instance, image_meta, [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] vm_ref = self.build_virtual_machine(instance, [ 651.171170] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] for vif in network_info: [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] return self._sync_wrapper(fn, *args, **kwargs) [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self.wait() [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self[:] = self._gt.wait() [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] return self._exit_event.wait() [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 651.171549] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] current.throw(*self._exc) [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] result = function(*args, **kwargs) [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] return func(*args, **kwargs) [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] raise e [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] nwinfo = self.network_api.allocate_for_instance( [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] created_port_ids = self._update_ports_for_instance( [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] with excutils.save_and_reraise_exception(): [ 651.171951] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self.force_reraise() [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] raise self.value [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] updated_port = self._update_port( [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] _ensure_no_port_binding_failure(port) [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] raise exception.PortBindingFailed(port_id=port['id']) [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] nova.exception.PortBindingFailed: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. [ 651.172308] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] [ 651.172308] env[62235]: INFO nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Terminating instance [ 651.173772] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Acquiring lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.323530] env[62235]: DEBUG nova.network.neutron [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.408350] env[62235]: DEBUG nova.network.neutron [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.477185] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.542218] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.542351] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 651.545374] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.748s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.546757] env[62235]: INFO nova.compute.claims [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.911120] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef781652-529a-413f-849a-d787d7b5c64f req-73433eee-bd2f-43e6-9fc5-942890e763ac service nova] Releasing lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.911595] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Acquired lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.911794] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.053740] env[62235]: DEBUG nova.compute.utils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.054962] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.055154] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.108354] env[62235]: DEBUG nova.policy [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d7ef1bc26534632ad28710f9df9803c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e22686dbba4420dad98d5de1f4fd449', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.412839] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Successfully created port: 639c8006-d55f-4601-895e-987463ed19b3 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.437133] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.561731] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 652.569615] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.874763] env[62235]: DEBUG nova.compute.manager [req-748c8a29-a0cf-422e-96e7-49f1f26294f1 req-5f0013f2-918f-4394-a1e5-7460304cc013 service nova] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Received event network-vif-deleted-8abc4405-49b4-4cc9-adfc-ff24ba7c3676 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.959265] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0997d4e-741a-4cae-97a1-4460bf2e05f9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.968912] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0692221a-6d75-410b-826a-ee399613b966 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.000780] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb75bf4-a81e-4c64-9c6f-a01a2d5fb351 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.008249] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8e4e14-6920-4152-8974-e453487f7549 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.023110] env[62235]: DEBUG nova.compute.provider_tree [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.074212] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Releasing lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.074597] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.074798] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 653.075095] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf6582a0-cd33-4303-ad07-50708a2e851f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.084975] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbec6d6-d1e2-4368-a4a2-e7340e137c2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.105259] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 60f868e1-e138-47d4-a510-5d3cd412e7de could not be found. [ 653.105501] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.105685] env[62235]: INFO nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Took 0.03 seconds to destroy the instance on the hypervisor. [ 653.105934] env[62235]: DEBUG oslo.service.loopingcall [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.106167] env[62235]: DEBUG nova.compute.manager [-] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.106261] env[62235]: DEBUG nova.network.neutron [-] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.133063] env[62235]: DEBUG nova.network.neutron [-] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.527514] env[62235]: DEBUG nova.scheduler.client.report [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.573392] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 653.599591] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.599831] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.599986] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.600247] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.600359] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.600503] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.600707] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.600861] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.601077] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.601297] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.601495] env[62235]: DEBUG nova.virt.hardware [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.602858] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc96309-2e6d-45ab-a857-0e91186e4975 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.611200] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3ebb9f-ab50-4708-bda0-af9eca79d0e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.632680] env[62235]: ERROR nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. [ 653.632680] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.632680] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.632680] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.632680] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.632680] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.632680] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.632680] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.632680] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.632680] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 653.632680] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.632680] env[62235]: ERROR nova.compute.manager raise self.value [ 653.632680] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.632680] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.632680] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.632680] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.633205] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.633205] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.633205] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. [ 653.633205] env[62235]: ERROR nova.compute.manager [ 653.633205] env[62235]: Traceback (most recent call last): [ 653.633205] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.633205] env[62235]: listener.cb(fileno) [ 653.633205] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.633205] env[62235]: result = function(*args, **kwargs) [ 653.633205] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.633205] env[62235]: return func(*args, **kwargs) [ 653.633205] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.633205] env[62235]: raise e [ 653.633205] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.633205] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 653.633205] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.633205] env[62235]: created_port_ids = self._update_ports_for_instance( [ 653.633205] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.633205] env[62235]: with excutils.save_and_reraise_exception(): [ 653.633205] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.633205] env[62235]: self.force_reraise() [ 653.633205] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.633205] env[62235]: raise self.value [ 653.633205] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.633205] env[62235]: updated_port = self._update_port( [ 653.633205] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.633205] env[62235]: _ensure_no_port_binding_failure(port) [ 653.633205] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.633205] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.633989] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. [ 653.633989] env[62235]: Removing descriptor: 16 [ 653.633989] env[62235]: ERROR nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Traceback (most recent call last): [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] yield resources [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self.driver.spawn(context, instance, image_meta, [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.633989] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] vm_ref = self.build_virtual_machine(instance, [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] for vif in network_info: [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return self._sync_wrapper(fn, *args, **kwargs) [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self.wait() [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self[:] = self._gt.wait() [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return self._exit_event.wait() [ 653.634372] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] result = hub.switch() [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return self.greenlet.switch() [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] result = function(*args, **kwargs) [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return func(*args, **kwargs) [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] raise e [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] nwinfo = self.network_api.allocate_for_instance( [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.634718] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] created_port_ids = self._update_ports_for_instance( [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] with excutils.save_and_reraise_exception(): [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self.force_reraise() [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] raise self.value [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] updated_port = self._update_port( [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] _ensure_no_port_binding_failure(port) [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.635094] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] raise exception.PortBindingFailed(port_id=port['id']) [ 653.635420] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] nova.exception.PortBindingFailed: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. [ 653.635420] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] [ 653.635420] env[62235]: INFO nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Terminating instance [ 653.635800] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.635915] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.636096] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.637503] env[62235]: DEBUG nova.network.neutron [-] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.036087] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.036652] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.039355] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.313s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.142107] env[62235]: INFO nova.compute.manager [-] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Took 1.04 seconds to deallocate network for instance. [ 654.144582] env[62235]: DEBUG nova.compute.claims [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.144755] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.154715] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.239646] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.544327] env[62235]: DEBUG nova.compute.utils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 654.545735] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 654.545908] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 654.600210] env[62235]: DEBUG nova.policy [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77482d78c3054bb1b015d6933e92712a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83d9b9725ec740b1bc994e7c39999935', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.743749] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.743749] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 654.743915] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 654.744227] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a65ee41b-3ce7-4d0f-9b90-7a94be84a78b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.759858] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425512ed-fd6b-4ddc-a450-4a8c8e5ba871 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.781711] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c4843d9-333f-4a08-9ec1-7971ecf53875 could not be found. [ 654.781942] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 654.782434] env[62235]: INFO nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Took 0.04 seconds to destroy the instance on the hypervisor. [ 654.782499] env[62235]: DEBUG oslo.service.loopingcall [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 654.785041] env[62235]: DEBUG nova.compute.manager [-] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.785134] env[62235]: DEBUG nova.network.neutron [-] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.802116] env[62235]: DEBUG nova.network.neutron [-] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.924404] env[62235]: DEBUG nova.compute.manager [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Received event network-changed-639c8006-d55f-4601-895e-987463ed19b3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.925143] env[62235]: DEBUG nova.compute.manager [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Refreshing instance network info cache due to event network-changed-639c8006-d55f-4601-895e-987463ed19b3. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 654.925143] env[62235]: DEBUG oslo_concurrency.lockutils [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] Acquiring lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.925143] env[62235]: DEBUG oslo_concurrency.lockutils [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] Acquired lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.925143] env[62235]: DEBUG nova.network.neutron [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Refreshing network info cache for port 639c8006-d55f-4601-895e-987463ed19b3 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.007192] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Successfully created port: 1f3048e6-62b0-4ff1-95e9-a1b04c04922f {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.010720] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8645b733-3e62-44dd-b6a4-60d833435fd7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.018623] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d22191-0238-42d0-ad26-b565112fa498 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.053484] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.065047] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad072c78-cd23-4f7e-ad9f-9618a5f4a900 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.070343] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61b52fe-aac4-4e5d-a33d-8a7800930d9c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.084335] env[62235]: DEBUG nova.compute.provider_tree [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.306406] env[62235]: DEBUG nova.network.neutron [-] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.455196] env[62235]: DEBUG nova.network.neutron [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.590655] env[62235]: DEBUG nova.scheduler.client.report [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.755795] env[62235]: DEBUG nova.network.neutron [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.808828] env[62235]: INFO nova.compute.manager [-] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Took 1.02 seconds to deallocate network for instance. [ 655.811744] env[62235]: DEBUG nova.compute.claims [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 655.811744] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.063703] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.090310] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.090564] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.090704] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.090865] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.091016] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.091487] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.092255] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.092255] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.092255] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.092407] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.092621] env[62235]: DEBUG nova.virt.hardware [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.093498] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a6b01d-bc1d-4503-b4a5-25093a4dbd1f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.097742] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.058s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.098590] env[62235]: ERROR nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Traceback (most recent call last): [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self.driver.spawn(context, instance, image_meta, [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] vm_ref = self.build_virtual_machine(instance, [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.098590] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] for vif in network_info: [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return self._sync_wrapper(fn, *args, **kwargs) [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self.wait() [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self[:] = self._gt.wait() [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return self._exit_event.wait() [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] result = hub.switch() [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.098964] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return self.greenlet.switch() [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] result = function(*args, **kwargs) [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] return func(*args, **kwargs) [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] raise e [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] nwinfo = self.network_api.allocate_for_instance( [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] created_port_ids = self._update_ports_for_instance( [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] with excutils.save_and_reraise_exception(): [ 656.099675] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] self.force_reraise() [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] raise self.value [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] updated_port = self._update_port( [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] _ensure_no_port_binding_failure(port) [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] raise exception.PortBindingFailed(port_id=port['id']) [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] nova.exception.PortBindingFailed: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. [ 656.100473] env[62235]: ERROR nova.compute.manager [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] [ 656.101586] env[62235]: DEBUG nova.compute.utils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.101586] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.577s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.104937] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Build of instance 74d069ed-c262-4b91-bcdd-ea3b4121adc4 was re-scheduled: Binding failed for port 3398a119-54e5-4d5a-b1b3-c91742140291, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.105430] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.105659] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Acquiring lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.105803] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Acquired lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.107082] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.111250] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb652d5-ec5d-4568-a5b4-b7a71bba7e64 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.258120] env[62235]: DEBUG oslo_concurrency.lockutils [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] Releasing lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.258424] env[62235]: DEBUG nova.compute.manager [req-dbda5ab5-db10-4ff5-8dc0-44825bd7b64a req-707765f8-1519-4391-9af2-3f7917c1bb19 service nova] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Received event network-vif-deleted-639c8006-d55f-4601-895e-987463ed19b3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.268631] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.268863] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.354652] env[62235]: ERROR nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. [ 656.354652] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.354652] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.354652] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.354652] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.354652] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.354652] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.354652] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.354652] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.354652] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 656.354652] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.354652] env[62235]: ERROR nova.compute.manager raise self.value [ 656.354652] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.354652] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.354652] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.354652] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.355476] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.355476] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.355476] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. [ 656.355476] env[62235]: ERROR nova.compute.manager [ 656.355476] env[62235]: Traceback (most recent call last): [ 656.355476] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.355476] env[62235]: listener.cb(fileno) [ 656.355476] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.355476] env[62235]: result = function(*args, **kwargs) [ 656.355476] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.355476] env[62235]: return func(*args, **kwargs) [ 656.355476] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.355476] env[62235]: raise e [ 656.355476] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.355476] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 656.355476] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.355476] env[62235]: created_port_ids = self._update_ports_for_instance( [ 656.355476] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.355476] env[62235]: with excutils.save_and_reraise_exception(): [ 656.355476] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.355476] env[62235]: self.force_reraise() [ 656.355476] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.355476] env[62235]: raise self.value [ 656.355476] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.355476] env[62235]: updated_port = self._update_port( [ 656.355476] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.355476] env[62235]: _ensure_no_port_binding_failure(port) [ 656.355476] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.355476] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.356407] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. [ 656.356407] env[62235]: Removing descriptor: 16 [ 656.356407] env[62235]: ERROR nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Traceback (most recent call last): [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] yield resources [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self.driver.spawn(context, instance, image_meta, [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.356407] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] vm_ref = self.build_virtual_machine(instance, [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] for vif in network_info: [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return self._sync_wrapper(fn, *args, **kwargs) [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self.wait() [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self[:] = self._gt.wait() [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return self._exit_event.wait() [ 656.356751] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] result = hub.switch() [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return self.greenlet.switch() [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] result = function(*args, **kwargs) [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return func(*args, **kwargs) [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] raise e [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] nwinfo = self.network_api.allocate_for_instance( [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.357124] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] created_port_ids = self._update_ports_for_instance( [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] with excutils.save_and_reraise_exception(): [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self.force_reraise() [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] raise self.value [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] updated_port = self._update_port( [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] _ensure_no_port_binding_failure(port) [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.357485] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] raise exception.PortBindingFailed(port_id=port['id']) [ 656.357857] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] nova.exception.PortBindingFailed: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. [ 656.357857] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] [ 656.357857] env[62235]: INFO nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Terminating instance [ 656.358087] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Acquiring lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.358313] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Acquired lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.359039] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.626946] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.704788] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.882836] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.945508] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.952329] env[62235]: DEBUG nova.compute.manager [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Received event network-changed-1f3048e6-62b0-4ff1-95e9-a1b04c04922f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.952329] env[62235]: DEBUG nova.compute.manager [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Refreshing instance network info cache due to event network-changed-1f3048e6-62b0-4ff1-95e9-a1b04c04922f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.952329] env[62235]: DEBUG oslo_concurrency.lockutils [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] Acquiring lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.021032] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480d4b17-bf02-4e54-ba1d-2f234692e321 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.037223] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159805d2-ba70-4ad6-a8d8-ae6e2fa42099 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.098367] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea585278-0adf-4264-909c-7b00febc745f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.111525] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efdd850-32cb-4628-be8a-b843948c94c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.134032] env[62235]: DEBUG nova.compute.provider_tree [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.207530] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Releasing lock "refresh_cache-74d069ed-c262-4b91-bcdd-ea3b4121adc4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.207793] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.207978] env[62235]: DEBUG nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.208164] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.231843] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.450840] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Releasing lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.451326] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 657.451514] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.451829] env[62235]: DEBUG oslo_concurrency.lockutils [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] Acquired lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.451992] env[62235]: DEBUG nova.network.neutron [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Refreshing network info cache for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 657.453249] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee8b40e8-751e-41be-9721-bf0aa548ac1a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.464062] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffbe3fb-fe25-437f-be44-897894fc0c9a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.484958] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f830c92d-ccf3-4612-980f-24ed8f679d4b could not be found. [ 657.485187] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.485458] env[62235]: INFO nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 657.485720] env[62235]: DEBUG oslo.service.loopingcall [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.485928] env[62235]: DEBUG nova.compute.manager [-] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.486080] env[62235]: DEBUG nova.network.neutron [-] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.508308] env[62235]: DEBUG nova.network.neutron [-] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.635707] env[62235]: DEBUG nova.scheduler.client.report [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.734752] env[62235]: DEBUG nova.network.neutron [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.974446] env[62235]: DEBUG nova.network.neutron [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.010040] env[62235]: DEBUG nova.network.neutron [-] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.028479] env[62235]: DEBUG nova.network.neutron [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.141545] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.041s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.142196] env[62235]: ERROR nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Traceback (most recent call last): [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self.driver.spawn(context, instance, image_meta, [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] vm_ref = self.build_virtual_machine(instance, [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.142196] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] for vif in network_info: [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return self._sync_wrapper(fn, *args, **kwargs) [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self.wait() [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self[:] = self._gt.wait() [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return self._exit_event.wait() [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] result = hub.switch() [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.142573] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return self.greenlet.switch() [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] result = function(*args, **kwargs) [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] return func(*args, **kwargs) [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] raise e [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] nwinfo = self.network_api.allocate_for_instance( [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] created_port_ids = self._update_ports_for_instance( [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] with excutils.save_and_reraise_exception(): [ 658.142968] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] self.force_reraise() [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] raise self.value [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] updated_port = self._update_port( [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] _ensure_no_port_binding_failure(port) [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] raise exception.PortBindingFailed(port_id=port['id']) [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] nova.exception.PortBindingFailed: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. [ 658.143369] env[62235]: ERROR nova.compute.manager [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] [ 658.143629] env[62235]: DEBUG nova.compute.utils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.144138] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.573s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.144320] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.144473] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 658.144747] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.292s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.147863] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Build of instance 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4 was re-scheduled: Binding failed for port f1d4d61e-8bb9-4979-abef-995e1da9c4a6, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 658.148278] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 658.148499] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquiring lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.148646] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Acquired lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.148803] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.150172] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c583f0-f61d-4299-a078-4f25da3cff1a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.159096] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e37c031-1b66-4aeb-a447-f293c235ff10 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.173823] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db27af5f-2fba-4424-a7f2-1774a3aad3a9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.180900] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231a98b9-0afd-4a96-ac3c-a0f6ec983f03 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.211210] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181554MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 658.211385] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.236873] env[62235]: INFO nova.compute.manager [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] [instance: 74d069ed-c262-4b91-bcdd-ea3b4121adc4] Took 1.03 seconds to deallocate network for instance. [ 658.513373] env[62235]: INFO nova.compute.manager [-] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Took 1.03 seconds to deallocate network for instance. [ 658.515711] env[62235]: DEBUG nova.compute.claims [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 658.515930] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.531489] env[62235]: DEBUG oslo_concurrency.lockutils [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] Releasing lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.531739] env[62235]: DEBUG nova.compute.manager [req-a339c5e1-31fe-485e-b373-ed4e9a9fd7c8 req-d4ca5da3-fae8-4de8-8ba6-6d4ee4a05aa9 service nova] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Received event network-vif-deleted-1f3048e6-62b0-4ff1-95e9-a1b04c04922f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.671753] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.715993] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.985586] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352ae6b0-ca80-418b-a09d-de333a3ea12e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.993502] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95488e0-68b1-496a-b170-e8f40dba3653 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.025554] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf80a44-5dc1-439e-bf66-51184d8a9420 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.036665] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b4cd49-3fe1-455b-80c7-3eb2b86c35fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.047428] env[62235]: DEBUG nova.compute.provider_tree [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.218294] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Releasing lock "refresh_cache-08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.218557] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 659.218722] env[62235]: DEBUG nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.218892] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.233833] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.265179] env[62235]: INFO nova.scheduler.client.report [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Deleted allocations for instance 74d069ed-c262-4b91-bcdd-ea3b4121adc4 [ 659.550874] env[62235]: DEBUG nova.scheduler.client.report [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.736216] env[62235]: DEBUG nova.network.neutron [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.775147] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7a03cf9c-ba4b-4be3-89cd-72c99e0a5976 tempest-ImagesNegativeTestJSON-335814720 tempest-ImagesNegativeTestJSON-335814720-project-member] Lock "74d069ed-c262-4b91-bcdd-ea3b4121adc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.221s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.056349] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.057013] env[62235]: ERROR nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Traceback (most recent call last): [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self.driver.spawn(context, instance, image_meta, [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] vm_ref = self.build_virtual_machine(instance, [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.057013] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] for vif in network_info: [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] return self._sync_wrapper(fn, *args, **kwargs) [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self.wait() [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self[:] = self._gt.wait() [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] return self._exit_event.wait() [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] current.throw(*self._exc) [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.057297] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] result = function(*args, **kwargs) [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] return func(*args, **kwargs) [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] raise e [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] nwinfo = self.network_api.allocate_for_instance( [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] created_port_ids = self._update_ports_for_instance( [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] with excutils.save_and_reraise_exception(): [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] self.force_reraise() [ 660.057611] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] raise self.value [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] updated_port = self._update_port( [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] _ensure_no_port_binding_failure(port) [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] raise exception.PortBindingFailed(port_id=port['id']) [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] nova.exception.PortBindingFailed: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. [ 660.057915] env[62235]: ERROR nova.compute.manager [instance: 91e05f87-96a4-4d52-9016-841a3a034930] [ 660.057915] env[62235]: DEBUG nova.compute.utils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.059106] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.734s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.063285] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Build of instance 91e05f87-96a4-4d52-9016-841a3a034930 was re-scheduled: Binding failed for port e230395f-db41-4e33-aa62-c1593806f605, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.063285] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.064455] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.064455] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquired lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.064455] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.239722] env[62235]: INFO nova.compute.manager [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] [instance: 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4] Took 1.02 seconds to deallocate network for instance. [ 660.277234] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.588339] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.737320] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.799938] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.897426] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c300b14c-b2e7-40c7-a31a-456f6b38325e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.904943] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3ab9b8-d7b1-41c5-9528-c01728f99bd8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.938807] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c668d0-64fa-46c3-aef4-fcdd5ccbee36 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.947563] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c958ae-7b3d-4adc-b10b-ae0ad4740cc1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.961304] env[62235]: DEBUG nova.compute.provider_tree [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.239585] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Releasing lock "refresh_cache-91e05f87-96a4-4d52-9016-841a3a034930" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.239833] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.240027] env[62235]: DEBUG nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.240201] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.266672] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.276237] env[62235]: INFO nova.scheduler.client.report [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Deleted allocations for instance 08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4 [ 661.465978] env[62235]: DEBUG nova.scheduler.client.report [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.770473] env[62235]: DEBUG nova.network.neutron [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.783274] env[62235]: DEBUG oslo_concurrency.lockutils [None req-765ea333-430b-41dd-8cef-b5eb10d46fce tempest-DeleteServersAdminTestJSON-487277677 tempest-DeleteServersAdminTestJSON-487277677-project-member] Lock "08baacd6-c807-4ad5-9d8e-0fdc8e4cb7b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.092s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.971036] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.972037] env[62235]: ERROR nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Traceback (most recent call last): [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self.driver.spawn(context, instance, image_meta, [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] vm_ref = self.build_virtual_machine(instance, [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.972037] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] for vif in network_info: [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] return self._sync_wrapper(fn, *args, **kwargs) [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self.wait() [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self[:] = self._gt.wait() [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] return self._exit_event.wait() [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] current.throw(*self._exc) [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.972382] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] result = function(*args, **kwargs) [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] return func(*args, **kwargs) [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] raise e [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] nwinfo = self.network_api.allocate_for_instance( [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] created_port_ids = self._update_ports_for_instance( [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] with excutils.save_and_reraise_exception(): [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] self.force_reraise() [ 661.972748] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] raise self.value [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] updated_port = self._update_port( [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] _ensure_no_port_binding_failure(port) [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] raise exception.PortBindingFailed(port_id=port['id']) [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] nova.exception.PortBindingFailed: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. [ 661.973115] env[62235]: ERROR nova.compute.manager [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] [ 661.973115] env[62235]: DEBUG nova.compute.utils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.973930] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.008s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.977469] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Build of instance 279097d2-348d-4f58-88ec-8e2e3143f754 was re-scheduled: Binding failed for port b8a5a8db-3cae-42a6-a67b-2c42b40ae081, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.978019] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.978253] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquiring lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.978403] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Acquired lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.978565] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.274216] env[62235]: INFO nova.compute.manager [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 91e05f87-96a4-4d52-9016-841a3a034930] Took 1.03 seconds to deallocate network for instance. [ 662.285630] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 662.502183] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.598409] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.803130] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.897148] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0cf585b-eba5-42c4-842f-c81cb9d85f95 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.906125] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb14f14-d200-4776-9147-b0ebef14a0ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.936024] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33f17f3-f896-4ecb-8467-7d452f6e69c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.943347] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006e9ed2-7ff1-4ab5-bb51-4ea16b3d707d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.956738] env[62235]: DEBUG nova.compute.provider_tree [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.101041] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Releasing lock "refresh_cache-279097d2-348d-4f58-88ec-8e2e3143f754" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.101487] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 663.101671] env[62235]: DEBUG nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.101851] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.127147] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.304077] env[62235]: INFO nova.scheduler.client.report [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Deleted allocations for instance 91e05f87-96a4-4d52-9016-841a3a034930 [ 663.462025] env[62235]: DEBUG nova.scheduler.client.report [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.630432] env[62235]: DEBUG nova.network.neutron [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.812755] env[62235]: DEBUG oslo_concurrency.lockutils [None req-72a06116-f66d-4686-8e5e-3da601126c46 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "91e05f87-96a4-4d52-9016-841a3a034930" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.434s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.968150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.968150] env[62235]: ERROR nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. [ 663.968150] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Traceback (most recent call last): [ 663.968150] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.968150] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self.driver.spawn(context, instance, image_meta, [ 663.968150] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.968150] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.968150] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.968150] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] vm_ref = self.build_virtual_machine(instance, [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] for vif in network_info: [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] return self._sync_wrapper(fn, *args, **kwargs) [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self.wait() [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self[:] = self._gt.wait() [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] return self._exit_event.wait() [ 663.968412] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] current.throw(*self._exc) [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] result = function(*args, **kwargs) [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] return func(*args, **kwargs) [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] raise e [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] nwinfo = self.network_api.allocate_for_instance( [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] created_port_ids = self._update_ports_for_instance( [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.968712] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] with excutils.save_and_reraise_exception(): [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] self.force_reraise() [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] raise self.value [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] updated_port = self._update_port( [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] _ensure_no_port_binding_failure(port) [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] raise exception.PortBindingFailed(port_id=port['id']) [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] nova.exception.PortBindingFailed: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. [ 663.969049] env[62235]: ERROR nova.compute.manager [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] [ 663.969421] env[62235]: DEBUG nova.compute.utils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.969421] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.534s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.970600] env[62235]: INFO nova.compute.claims [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.973405] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Build of instance 1d1a57fc-6bc7-47a1-918e-8d23ac797d54 was re-scheduled: Binding failed for port 58b8b07e-114d-40bb-a937-07d349e00319, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.973960] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.974217] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Acquiring lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.974368] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Acquired lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.974528] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.137882] env[62235]: INFO nova.compute.manager [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] [instance: 279097d2-348d-4f58-88ec-8e2e3143f754] Took 1.04 seconds to deallocate network for instance. [ 664.317195] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.500598] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.650452] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.851498] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.159150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Releasing lock "refresh_cache-1d1a57fc-6bc7-47a1-918e-8d23ac797d54" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.159436] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 665.159614] env[62235]: DEBUG nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.159915] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.171729] env[62235]: INFO nova.scheduler.client.report [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Deleted allocations for instance 279097d2-348d-4f58-88ec-8e2e3143f754 [ 665.200250] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.396313] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb1e984-e082-47b1-a257-b914c78dc716 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.403774] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa915cef-756b-451c-ae2e-1dede7eb62ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.441561] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe4c9ca-5c59-4a7d-a658-fb94aeb7e45a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.450498] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0df9a5-9669-4ebc-aed0-8cad924aeee7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.465654] env[62235]: DEBUG nova.compute.provider_tree [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.679634] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0cad2b81-f987-4b88-b590-f6d2bf586943 tempest-ServerRescueNegativeTestJSON-1573523070 tempest-ServerRescueNegativeTestJSON-1573523070-project-member] Lock "279097d2-348d-4f58-88ec-8e2e3143f754" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.619s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.703315] env[62235]: DEBUG nova.network.neutron [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.968869] env[62235]: DEBUG nova.scheduler.client.report [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.188838] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 666.208513] env[62235]: INFO nova.compute.manager [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] [instance: 1d1a57fc-6bc7-47a1-918e-8d23ac797d54] Took 1.05 seconds to deallocate network for instance. [ 666.254582] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.254875] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.475802] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.475802] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.477708] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.479257] env[62235]: INFO nova.compute.claims [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.717346] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.987414] env[62235]: DEBUG nova.compute.utils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.988909] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.989134] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 667.063525] env[62235]: DEBUG nova.policy [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fce2bb79c4014131afda03401cd281ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1715053dd59b475fbe88cf8b9e191ba0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.246919] env[62235]: INFO nova.scheduler.client.report [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Deleted allocations for instance 1d1a57fc-6bc7-47a1-918e-8d23ac797d54 [ 667.380175] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Successfully created port: c5eb8ae5-46e9-4185-8194-9ee04583ef3f {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.492365] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.760911] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f381136b-49ef-4b28-9fb0-2465871b2375 tempest-InstanceActionsNegativeTestJSON-1501329333 tempest-InstanceActionsNegativeTestJSON-1501329333-project-member] Lock "1d1a57fc-6bc7-47a1-918e-8d23ac797d54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.671s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.888170] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6602e94-6a8e-4b9b-91c8-b1a4f984a5ca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.896989] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9baa106-3de9-4ee3-b45e-d2b189b799e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.933976] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54f07ae-cc64-445d-9751-aba2e2b2dd0d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.941713] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b9cb95-dd5b-48a6-abaa-3491a396d302 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.958414] env[62235]: DEBUG nova.compute.provider_tree [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.267222] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.462242] env[62235]: DEBUG nova.scheduler.client.report [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.506804] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.538782] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.539063] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.539298] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.539448] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.539682] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.539838] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.540322] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.540501] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.540681] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.540856] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.541319] env[62235]: DEBUG nova.virt.hardware [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.542194] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12016d36-de48-4299-894f-95740e84a6ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.555589] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207daf2d-fb02-471b-b93d-1d213edcdb82 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.794343] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.972201] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.972849] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.977959] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.833s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.076534] env[62235]: DEBUG nova.compute.manager [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Received event network-changed-c5eb8ae5-46e9-4185-8194-9ee04583ef3f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.076654] env[62235]: DEBUG nova.compute.manager [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Refreshing instance network info cache due to event network-changed-c5eb8ae5-46e9-4185-8194-9ee04583ef3f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.076858] env[62235]: DEBUG oslo_concurrency.lockutils [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] Acquiring lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.077014] env[62235]: DEBUG oslo_concurrency.lockutils [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] Acquired lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.077181] env[62235]: DEBUG nova.network.neutron [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Refreshing network info cache for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.485446] env[62235]: DEBUG nova.compute.utils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.490126] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.490126] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 669.549889] env[62235]: ERROR nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. [ 669.549889] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.549889] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.549889] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.549889] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.549889] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.549889] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.549889] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.549889] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.549889] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 669.549889] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.549889] env[62235]: ERROR nova.compute.manager raise self.value [ 669.549889] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.549889] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.549889] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.549889] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.550400] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.550400] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.550400] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. [ 669.550400] env[62235]: ERROR nova.compute.manager [ 669.550400] env[62235]: Traceback (most recent call last): [ 669.550400] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.550400] env[62235]: listener.cb(fileno) [ 669.550400] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.550400] env[62235]: result = function(*args, **kwargs) [ 669.550400] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.550400] env[62235]: return func(*args, **kwargs) [ 669.550400] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.550400] env[62235]: raise e [ 669.550400] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.550400] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 669.550400] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.550400] env[62235]: created_port_ids = self._update_ports_for_instance( [ 669.550400] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.550400] env[62235]: with excutils.save_and_reraise_exception(): [ 669.550400] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.550400] env[62235]: self.force_reraise() [ 669.550400] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.550400] env[62235]: raise self.value [ 669.550400] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.550400] env[62235]: updated_port = self._update_port( [ 669.550400] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.550400] env[62235]: _ensure_no_port_binding_failure(port) [ 669.550400] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.550400] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.551172] env[62235]: nova.exception.PortBindingFailed: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. [ 669.551172] env[62235]: Removing descriptor: 16 [ 669.552774] env[62235]: ERROR nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Traceback (most recent call last): [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] yield resources [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self.driver.spawn(context, instance, image_meta, [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] vm_ref = self.build_virtual_machine(instance, [ 669.552774] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] for vif in network_info: [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return self._sync_wrapper(fn, *args, **kwargs) [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self.wait() [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self[:] = self._gt.wait() [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return self._exit_event.wait() [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.553271] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] result = hub.switch() [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return self.greenlet.switch() [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] result = function(*args, **kwargs) [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return func(*args, **kwargs) [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] raise e [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] nwinfo = self.network_api.allocate_for_instance( [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] created_port_ids = self._update_ports_for_instance( [ 669.553668] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] with excutils.save_and_reraise_exception(): [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self.force_reraise() [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] raise self.value [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] updated_port = self._update_port( [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] _ensure_no_port_binding_failure(port) [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] raise exception.PortBindingFailed(port_id=port['id']) [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] nova.exception.PortBindingFailed: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. [ 669.553991] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] [ 669.554524] env[62235]: INFO nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Terminating instance [ 669.554524] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Acquiring lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.614632] env[62235]: DEBUG nova.network.neutron [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.644915] env[62235]: DEBUG nova.policy [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 669.894286] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131c2a31-4210-4457-b873-15e096e47e2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.900999] env[62235]: DEBUG nova.network.neutron [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.907830] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606f937b-b8ba-46ca-95da-3a8fc0e2d071 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.943559] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b7ad7d-457e-4c74-938e-f090cc55b67b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.951038] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ab70fa-e46b-4d18-82c3-b9c6d0728219 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.967920] env[62235]: DEBUG nova.compute.provider_tree [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.990397] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 670.404640] env[62235]: DEBUG oslo_concurrency.lockutils [req-208ec067-ac00-417d-9597-02a00d272a6d req-15eda91d-b53c-4209-9936-416a58f1b5fb service nova] Releasing lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.405447] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Acquired lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.405942] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.471048] env[62235]: DEBUG nova.scheduler.client.report [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.506844] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "ec1830c7-7095-45d1-a188-c2993c60035e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.507146] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "ec1830c7-7095-45d1-a188-c2993c60035e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.549148] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Successfully created port: c3347862-c63a-4a4f-9f97-ca812573ba26 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.869247] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.869522] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.941421] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.980074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.980074] env[62235]: ERROR nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. [ 670.980074] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Traceback (most recent call last): [ 670.980074] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.980074] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self.driver.spawn(context, instance, image_meta, [ 670.980074] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.980074] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.980074] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.980074] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] vm_ref = self.build_virtual_machine(instance, [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] for vif in network_info: [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] return self._sync_wrapper(fn, *args, **kwargs) [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self.wait() [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self[:] = self._gt.wait() [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] return self._exit_event.wait() [ 670.981287] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] current.throw(*self._exc) [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] result = function(*args, **kwargs) [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] return func(*args, **kwargs) [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] raise e [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] nwinfo = self.network_api.allocate_for_instance( [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] created_port_ids = self._update_ports_for_instance( [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.981755] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] with excutils.save_and_reraise_exception(): [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] self.force_reraise() [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] raise self.value [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] updated_port = self._update_port( [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] _ensure_no_port_binding_failure(port) [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] raise exception.PortBindingFailed(port_id=port['id']) [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] nova.exception.PortBindingFailed: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. [ 670.982124] env[62235]: ERROR nova.compute.manager [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] [ 670.982463] env[62235]: DEBUG nova.compute.utils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.983784] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.168s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.987732] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Build of instance 60f868e1-e138-47d4-a510-5d3cd412e7de was re-scheduled: Binding failed for port 8abc4405-49b4-4cc9-adfc-ff24ba7c3676, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.988907] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.989399] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Acquiring lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.989653] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Acquired lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.989983] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.003157] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 671.037922] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.038171] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.038907] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.038907] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.038907] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.038907] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.039074] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.039298] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.039466] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.039622] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.039787] env[62235]: DEBUG nova.virt.hardware [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.041430] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4efc60c-f19a-4284-87cb-d3f211c3964c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.050483] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b34cd86-043e-48e5-81c7-c24a1bb9c254 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.111688] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.352198] env[62235]: DEBUG nova.compute.manager [req-fdf7adc0-b316-4af7-a9fb-3a7633824021 req-17f97008-3b9e-4ce5-8bfa-35955036cd3e service nova] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Received event network-vif-deleted-c5eb8ae5-46e9-4185-8194-9ee04583ef3f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.539271] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.623455] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Releasing lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.623455] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.623455] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.623455] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac2d611e-15d1-4524-b8db-25313214f9cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.632386] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c11c4c5-cfff-4bd8-82ce-3841a0c51fa2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.666723] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 584b41c1-976d-49e9-b5ea-c5c22e1877c5 could not be found. [ 671.666963] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.667177] env[62235]: INFO nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 671.667509] env[62235]: DEBUG oslo.service.loopingcall [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.670460] env[62235]: DEBUG nova.compute.manager [-] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.670560] env[62235]: DEBUG nova.network.neutron [-] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.693897] env[62235]: DEBUG nova.network.neutron [-] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.758458] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.985471] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a25e87-096f-4eb3-84b6-d648e648c9cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.997780] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c593cb6-a781-4b41-9faf-b47bc0d7e6f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.030832] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701355b5-f952-4b7d-8548-b2c37da68121 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.038148] env[62235]: ERROR nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. [ 672.038148] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.038148] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.038148] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.038148] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.038148] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.038148] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.038148] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.038148] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.038148] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 672.038148] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.038148] env[62235]: ERROR nova.compute.manager raise self.value [ 672.038148] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.038148] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.038148] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.038148] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.038662] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.038662] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.038662] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. [ 672.038662] env[62235]: ERROR nova.compute.manager [ 672.038662] env[62235]: Traceback (most recent call last): [ 672.038662] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.038662] env[62235]: listener.cb(fileno) [ 672.038662] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.038662] env[62235]: result = function(*args, **kwargs) [ 672.038662] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.038662] env[62235]: return func(*args, **kwargs) [ 672.038662] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.038662] env[62235]: raise e [ 672.038662] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.038662] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 672.038662] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.038662] env[62235]: created_port_ids = self._update_ports_for_instance( [ 672.038662] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.038662] env[62235]: with excutils.save_and_reraise_exception(): [ 672.038662] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.038662] env[62235]: self.force_reraise() [ 672.038662] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.038662] env[62235]: raise self.value [ 672.038662] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.038662] env[62235]: updated_port = self._update_port( [ 672.038662] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.038662] env[62235]: _ensure_no_port_binding_failure(port) [ 672.038662] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.038662] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.039936] env[62235]: nova.exception.PortBindingFailed: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. [ 672.039936] env[62235]: Removing descriptor: 22 [ 672.039936] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70daec93-0484-4ebc-8b6a-102c0d81cfa7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.044562] env[62235]: ERROR nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Traceback (most recent call last): [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] yield resources [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self.driver.spawn(context, instance, image_meta, [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] vm_ref = self.build_virtual_machine(instance, [ 672.044562] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] for vif in network_info: [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return self._sync_wrapper(fn, *args, **kwargs) [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self.wait() [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self[:] = self._gt.wait() [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return self._exit_event.wait() [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.044835] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] result = hub.switch() [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return self.greenlet.switch() [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] result = function(*args, **kwargs) [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return func(*args, **kwargs) [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] raise e [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] nwinfo = self.network_api.allocate_for_instance( [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] created_port_ids = self._update_ports_for_instance( [ 672.045278] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] with excutils.save_and_reraise_exception(): [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self.force_reraise() [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] raise self.value [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] updated_port = self._update_port( [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] _ensure_no_port_binding_failure(port) [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] raise exception.PortBindingFailed(port_id=port['id']) [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] nova.exception.PortBindingFailed: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. [ 672.045593] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] [ 672.045901] env[62235]: INFO nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Terminating instance [ 672.047724] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.047920] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.048263] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.058768] env[62235]: DEBUG nova.compute.provider_tree [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.080221] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.186227] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.196561] env[62235]: DEBUG nova.network.neutron [-] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.260815] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Releasing lock "refresh_cache-60f868e1-e138-47d4-a510-5d3cd412e7de" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.260911] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.262040] env[62235]: DEBUG nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.262040] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.275870] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.564748] env[62235]: DEBUG nova.scheduler.client.report [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.689076] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.689538] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.689734] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 672.690048] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74e48de8-c5d0-4416-b86c-7bdbb5e8f1b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.778196] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a576a33c-a6f0-4ee0-a0eb-8ce185078ce2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.778196] env[62235]: INFO nova.compute.manager [-] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Took 1.04 seconds to deallocate network for instance. [ 672.778196] env[62235]: DEBUG nova.compute.claims [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.778196] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.778196] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4d10be8a-a4ba-46f7-b4da-c37f8520bab6 could not be found. [ 672.778196] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.780300] env[62235]: INFO nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 672.780300] env[62235]: DEBUG oslo.service.loopingcall [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.780300] env[62235]: DEBUG nova.compute.manager [-] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.780300] env[62235]: DEBUG nova.network.neutron [-] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.780300] env[62235]: DEBUG nova.network.neutron [-] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.781016] env[62235]: DEBUG nova.network.neutron [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.069223] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.089s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.069940] env[62235]: ERROR nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Traceback (most recent call last): [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self.driver.spawn(context, instance, image_meta, [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] vm_ref = self.build_virtual_machine(instance, [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.069940] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] for vif in network_info: [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return self._sync_wrapper(fn, *args, **kwargs) [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self.wait() [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self[:] = self._gt.wait() [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return self._exit_event.wait() [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] result = hub.switch() [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 673.070262] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return self.greenlet.switch() [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] result = function(*args, **kwargs) [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] return func(*args, **kwargs) [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] raise e [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] nwinfo = self.network_api.allocate_for_instance( [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] created_port_ids = self._update_ports_for_instance( [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] with excutils.save_and_reraise_exception(): [ 673.070668] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] self.force_reraise() [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] raise self.value [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] updated_port = self._update_port( [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] _ensure_no_port_binding_failure(port) [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] raise exception.PortBindingFailed(port_id=port['id']) [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] nova.exception.PortBindingFailed: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. [ 673.071023] env[62235]: ERROR nova.compute.manager [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] [ 673.071369] env[62235]: DEBUG nova.compute.utils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.075027] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Build of instance 8c4843d9-333f-4a08-9ec1-7971ecf53875 was re-scheduled: Binding failed for port 639c8006-d55f-4601-895e-987463ed19b3, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.075027] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.075027] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.075027] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.075027] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.075588] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.863s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.273430] env[62235]: DEBUG nova.network.neutron [-] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.283656] env[62235]: INFO nova.compute.manager [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] [instance: 60f868e1-e138-47d4-a510-5d3cd412e7de] Took 1.02 seconds to deallocate network for instance. [ 673.456023] env[62235]: DEBUG nova.compute.manager [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Received event network-changed-c3347862-c63a-4a4f-9f97-ca812573ba26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.456023] env[62235]: DEBUG nova.compute.manager [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Refreshing instance network info cache due to event network-changed-c3347862-c63a-4a4f-9f97-ca812573ba26. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 673.456023] env[62235]: DEBUG oslo_concurrency.lockutils [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] Acquiring lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.456023] env[62235]: DEBUG oslo_concurrency.lockutils [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] Acquired lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.456023] env[62235]: DEBUG nova.network.neutron [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Refreshing network info cache for port c3347862-c63a-4a4f-9f97-ca812573ba26 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 673.599797] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.691970] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.776532] env[62235]: INFO nova.compute.manager [-] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Took 1.04 seconds to deallocate network for instance. [ 673.778759] env[62235]: DEBUG nova.compute.claims [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.778936] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.974732] env[62235]: DEBUG nova.network.neutron [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.068128] env[62235]: DEBUG nova.network.neutron [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.197111] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "refresh_cache-8c4843d9-333f-4a08-9ec1-7971ecf53875" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.197357] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.197544] env[62235]: DEBUG nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.197712] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.214515] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.328025] env[62235]: INFO nova.scheduler.client.report [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Deleted allocations for instance 60f868e1-e138-47d4-a510-5d3cd412e7de [ 674.573700] env[62235]: DEBUG oslo_concurrency.lockutils [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] Releasing lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.573997] env[62235]: DEBUG nova.compute.manager [req-6bfc9e39-7fe4-4da8-8959-d86c21a8c13e req-4810e103-069d-4c9b-9549-c5bc433d7c4e service nova] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Received event network-vif-deleted-c3347862-c63a-4a4f-9f97-ca812573ba26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.619630] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 60f868e1-e138-47d4-a510-5d3cd412e7de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.723557] env[62235]: DEBUG nova.network.neutron [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.843904] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baecf2d7-3a11-413f-be6a-de2cbc69a773 tempest-ServerMetadataNegativeTestJSON-1254121032 tempest-ServerMetadataNegativeTestJSON-1254121032-project-member] Lock "60f868e1-e138-47d4-a510-5d3cd412e7de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.344s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.126413] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 8c4843d9-333f-4a08-9ec1-7971ecf53875 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.126584] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance f830c92d-ccf3-4612-980f-24ed8f679d4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 675.126711] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 584b41c1-976d-49e9-b5ea-c5c22e1877c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 675.126831] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 4d10be8a-a4ba-46f7-b4da-c37f8520bab6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 675.225589] env[62235]: INFO nova.compute.manager [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 8c4843d9-333f-4a08-9ec1-7971ecf53875] Took 1.03 seconds to deallocate network for instance. [ 675.348611] env[62235]: DEBUG nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.631153] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.885231] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.135821] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3f245678-41f4-4685-87fc-822bf877d43e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.270941] env[62235]: INFO nova.scheduler.client.report [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted allocations for instance 8c4843d9-333f-4a08-9ec1-7971ecf53875 [ 676.639076] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 19e40767-d9f5-4b78-80f0-7f0f24464d40 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.785795] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a44db1d8-98a8-41bb-b89f-084f722159c4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "8c4843d9-333f-4a08-9ec1-7971ecf53875" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.413s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.145570] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b348bc9f-3bd4-475f-9957-42a99848005c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.287553] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.655728] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 012dd63b-a092-4a1d-a79c-1e1bf2fd8771 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.814268] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.155323] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance be4fb088-8b01-4245-82fb-b0e085825597 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.253560] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.254175] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.658271] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 693db195-ce52-40ff-bbb4-91f03d5a0bd9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.913562] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.913989] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.166814] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 452cf04e-d6e9-48a8-a77b-f931c9440858 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.672625] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 1c6f4360-b80b-43e3-accc-f7500e626100 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.176013] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ea4a3f54-388e-451f-9318-78943a169213 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.214020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "590440de-4b3b-4019-b0c3-58b5ad329c88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.214325] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.682940] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3b4167dd-4535-4482-9452-5548e07dc581 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.184444] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 86e61430-ce9a-436e-9331-6276604610e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.688244] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 430056ca-ccba-45c5-adf7-9c068f81f9ec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.193493] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance c62172db-7e1c-4fd1-b8b9-a8daf0a71f91 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.698259] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 7e6ebba9-58e8-4971-b9fb-8365a61d5505 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.128216] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.128446] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.201587] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.705911] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 6671163c-c829-4046-97e5-e85fe5e223dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.212351] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3e7a81ae-2163-4562-b483-dc639c602fa0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.716646] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance dd7268d1-1da5-4e5f-9d45-816df147b564 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.222368] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance d6324b5d-ec70-4bb4-b3c6-b1524abd0034 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.725887] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance afa35ae0-40c6-4eaa-80f5-c1ada27f1871 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.230351] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ec1830c7-7095-45d1-a188-c2993c60035e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.734229] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b4a3777f-2ee5-464b-84eb-0722d59f2b44 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.734509] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 686.734546] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 687.097016] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd919f47-309b-4329-b003-ce37359df4ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.104487] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aff749f-28a7-41c2-9fa7-b66bc4c9586e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.133781] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49233b04-9a39-464b-93b4-5ba7e5809f74 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.140734] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686a281f-d782-4ab6-aa25-5256da7d9ca0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.153448] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.656843] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.162417] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 688.162690] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.088s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.162964] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.647s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.989223] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9164f2ab-264b-476b-aa15-129a4e5a3b54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.997347] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300461ed-61bb-472b-bde4-0af7928d7e04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.028058] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279a3f3e-af16-444a-afe2-46ec523c6450 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.035514] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c450e8d-3f9e-470a-be55-6d5efd1b4ffe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.048674] env[62235]: DEBUG nova.compute.provider_tree [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.551624] env[62235]: DEBUG nova.scheduler.client.report [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.056678] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.057085] env[62235]: ERROR nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Traceback (most recent call last): [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self.driver.spawn(context, instance, image_meta, [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] vm_ref = self.build_virtual_machine(instance, [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.057085] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] for vif in network_info: [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return self._sync_wrapper(fn, *args, **kwargs) [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self.wait() [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self[:] = self._gt.wait() [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return self._exit_event.wait() [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] result = hub.switch() [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.057425] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return self.greenlet.switch() [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] result = function(*args, **kwargs) [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] return func(*args, **kwargs) [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] raise e [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] nwinfo = self.network_api.allocate_for_instance( [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] created_port_ids = self._update_ports_for_instance( [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] with excutils.save_and_reraise_exception(): [ 690.057746] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] self.force_reraise() [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] raise self.value [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] updated_port = self._update_port( [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] _ensure_no_port_binding_failure(port) [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] raise exception.PortBindingFailed(port_id=port['id']) [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] nova.exception.PortBindingFailed: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. [ 690.058069] env[62235]: ERROR nova.compute.manager [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] [ 690.058352] env[62235]: DEBUG nova.compute.utils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.059675] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Build of instance f830c92d-ccf3-4612-980f-24ed8f679d4b was re-scheduled: Binding failed for port 1f3048e6-62b0-4ff1-95e9-a1b04c04922f, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.060108] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.060345] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Acquiring lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.060495] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Acquired lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.060651] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.062822] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.263s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.065306] env[62235]: INFO nova.compute.claims [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.587207] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.647950] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.152794] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Releasing lock "refresh_cache-f830c92d-ccf3-4612-980f-24ed8f679d4b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.153044] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.153250] env[62235]: DEBUG nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.153415] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.171083] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.418371] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f45bd97-f3ef-43b3-bd52-3c1ab8c83ab0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.426014] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1af4974-7fff-4d8c-b588-591de53032b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.460795] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986b3ef1-90da-4e90-975e-218c931b2734 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.465557] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3de7ec-25d6-4b24-931d-74345d43025f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.478215] env[62235]: DEBUG nova.compute.provider_tree [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.671406] env[62235]: DEBUG nova.network.neutron [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.981612] env[62235]: DEBUG nova.scheduler.client.report [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.174979] env[62235]: INFO nova.compute.manager [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] [instance: f830c92d-ccf3-4612-980f-24ed8f679d4b] Took 1.02 seconds to deallocate network for instance. [ 692.485903] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.486463] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.489740] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.687s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.491121] env[62235]: INFO nova.compute.claims [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.995191] env[62235]: DEBUG nova.compute.utils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.998233] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 692.998402] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 693.035033] env[62235]: DEBUG nova.policy [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84b3519ef1f04cd29ad7a9b70d9204ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f36c6cd46a8b4e59936b22cc7d409f37', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.203065] env[62235]: INFO nova.scheduler.client.report [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Deleted allocations for instance f830c92d-ccf3-4612-980f-24ed8f679d4b [ 693.396040] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Successfully created port: 35b9ca76-9b0e-4441-8212-f56017629258 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.504154] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.715734] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b1b65f33-5d7c-4db9-b4a3-3fb755a64b4a tempest-ServersAdminNegativeTestJSON-1918558665 tempest-ServersAdminNegativeTestJSON-1918558665-project-member] Lock "f830c92d-ccf3-4612-980f-24ed8f679d4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.059s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.939321] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a917983f-73d4-4e25-9628-09755574e9da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.946554] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368822f8-fe06-463f-8f59-352965e932d5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.979401] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc868c71-6d47-4f08-91a5-f765766042cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.988889] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4045e456-a790-4962-87f9-f251f42707b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.002740] env[62235]: DEBUG nova.compute.provider_tree [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.218110] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.255781] env[62235]: DEBUG nova.compute.manager [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Received event network-changed-35b9ca76-9b0e-4441-8212-f56017629258 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.256036] env[62235]: DEBUG nova.compute.manager [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Refreshing instance network info cache due to event network-changed-35b9ca76-9b0e-4441-8212-f56017629258. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 694.256448] env[62235]: DEBUG oslo_concurrency.lockutils [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] Acquiring lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.256550] env[62235]: DEBUG oslo_concurrency.lockutils [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] Acquired lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.257322] env[62235]: DEBUG nova.network.neutron [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Refreshing network info cache for port 35b9ca76-9b0e-4441-8212-f56017629258 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.437015] env[62235]: ERROR nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. [ 694.437015] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.437015] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.437015] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.437015] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.437015] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.437015] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.437015] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.437015] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.437015] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 694.437015] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.437015] env[62235]: ERROR nova.compute.manager raise self.value [ 694.437015] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.437015] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.437015] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.437015] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.437586] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.437586] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.437586] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. [ 694.437586] env[62235]: ERROR nova.compute.manager [ 694.437586] env[62235]: Traceback (most recent call last): [ 694.437586] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.437586] env[62235]: listener.cb(fileno) [ 694.437586] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.437586] env[62235]: result = function(*args, **kwargs) [ 694.437586] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.437586] env[62235]: return func(*args, **kwargs) [ 694.437586] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.437586] env[62235]: raise e [ 694.437586] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.437586] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 694.437586] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.437586] env[62235]: created_port_ids = self._update_ports_for_instance( [ 694.437586] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.437586] env[62235]: with excutils.save_and_reraise_exception(): [ 694.437586] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.437586] env[62235]: self.force_reraise() [ 694.437586] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.437586] env[62235]: raise self.value [ 694.437586] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.437586] env[62235]: updated_port = self._update_port( [ 694.437586] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.437586] env[62235]: _ensure_no_port_binding_failure(port) [ 694.437586] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.437586] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.438730] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. [ 694.438730] env[62235]: Removing descriptor: 22 [ 694.508560] env[62235]: DEBUG nova.scheduler.client.report [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.513079] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.537570] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.537700] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.537788] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.537971] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.538130] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.538350] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.538590] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.538682] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.538782] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.538944] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.539142] env[62235]: DEBUG nova.virt.hardware [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.540028] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b025195c-51ae-40e6-930c-8ac359beb8ff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.549957] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e93fba-ba3c-4bbe-92c5-6a34238172b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.562975] env[62235]: ERROR nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Traceback (most recent call last): [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] yield resources [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self.driver.spawn(context, instance, image_meta, [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] vm_ref = self.build_virtual_machine(instance, [ 694.562975] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] for vif in network_info: [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] return self._sync_wrapper(fn, *args, **kwargs) [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self.wait() [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self[:] = self._gt.wait() [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] return self._exit_event.wait() [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.563315] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] current.throw(*self._exc) [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] result = function(*args, **kwargs) [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] return func(*args, **kwargs) [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] raise e [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] nwinfo = self.network_api.allocate_for_instance( [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] created_port_ids = self._update_ports_for_instance( [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] with excutils.save_and_reraise_exception(): [ 694.567109] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self.force_reraise() [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] raise self.value [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] updated_port = self._update_port( [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] _ensure_no_port_binding_failure(port) [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] raise exception.PortBindingFailed(port_id=port['id']) [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] nova.exception.PortBindingFailed: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. [ 694.567556] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] [ 694.567556] env[62235]: INFO nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Terminating instance [ 694.567876] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Acquiring lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.742100] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.777305] env[62235]: DEBUG nova.network.neutron [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.893337] env[62235]: DEBUG nova.network.neutron [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.014020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.014580] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.017235] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.166s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.018929] env[62235]: INFO nova.compute.claims [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.395884] env[62235]: DEBUG oslo_concurrency.lockutils [req-74f240b3-9045-4dd6-838c-5bfc0c0070d5 req-b346e414-b426-404c-b755-420bf5121645 service nova] Releasing lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.396327] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Acquired lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.396514] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.526116] env[62235]: DEBUG nova.compute.utils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.528971] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.529472] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 695.602758] env[62235]: DEBUG nova.policy [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85d819ef90ee4aabbf4fe48e639beee6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c226dc4526f4455bf081061756b5727', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.930401] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.030174] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.278462] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.341826] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Successfully created port: a11d8a7d-c492-408a-95ac-49da87650c44 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.345961] env[62235]: DEBUG nova.compute.manager [req-65bd53e5-dbf8-4a1e-bc2f-e194866d0a8d req-38373193-82ee-485e-bf68-ee54c05d5ac1 service nova] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Received event network-vif-deleted-35b9ca76-9b0e-4441-8212-f56017629258 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.466332] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a03fa1e-d3b2-48bc-b149-f3f001202b41 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.476023] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4f84f4-20d9-4a70-9db9-805557f6e12b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.516193] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5de336-d124-4017-94a3-4941eb2bf715 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.522276] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e2c42c-cf77-4eb2-81ee-77d1f38d0339 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.537292] env[62235]: DEBUG nova.compute.provider_tree [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.782023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Releasing lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.785136] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.785136] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.785136] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41c9db69-ebc1-4a15-b5d8-c119dc1cf60d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.792669] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c5309b-a1f7-46bb-a568-806e71ca4274 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.815815] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2 could not be found. [ 696.816268] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.816610] env[62235]: INFO nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 696.817035] env[62235]: DEBUG oslo.service.loopingcall [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.817618] env[62235]: DEBUG nova.compute.manager [-] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.817850] env[62235]: DEBUG nova.network.neutron [-] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.849945] env[62235]: DEBUG nova.network.neutron [-] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.049187] env[62235]: DEBUG nova.scheduler.client.report [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.056095] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 697.088361] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:19:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d2035966-33b3-4adb-b834-275d8138b0cf',id=33,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-6869826',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.088902] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.089360] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.089712] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.091066] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.091066] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.091066] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.091066] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.091066] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.091225] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.091225] env[62235]: DEBUG nova.virt.hardware [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.092204] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25c30cf-23dd-48ee-928e-8926ed964c67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.102154] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2c7681-67ba-4e6d-904b-ce469cd1cfce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.351022] env[62235]: DEBUG nova.network.neutron [-] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.560865] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.562024] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.566621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.849s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.569800] env[62235]: INFO nova.compute.claims [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.853918] env[62235]: INFO nova.compute.manager [-] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Took 1.04 seconds to deallocate network for instance. [ 697.858063] env[62235]: DEBUG nova.compute.claims [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.858063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.018155] env[62235]: ERROR nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. [ 698.018155] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 698.018155] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.018155] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 698.018155] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.018155] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 698.018155] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.018155] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 698.018155] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.018155] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 698.018155] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.018155] env[62235]: ERROR nova.compute.manager raise self.value [ 698.018155] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.018155] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 698.018155] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.018155] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 698.018672] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.018672] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 698.018672] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. [ 698.018672] env[62235]: ERROR nova.compute.manager [ 698.018672] env[62235]: Traceback (most recent call last): [ 698.018672] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 698.018672] env[62235]: listener.cb(fileno) [ 698.018672] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.018672] env[62235]: result = function(*args, **kwargs) [ 698.018672] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.018672] env[62235]: return func(*args, **kwargs) [ 698.018672] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.018672] env[62235]: raise e [ 698.018672] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.018672] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 698.018672] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.018672] env[62235]: created_port_ids = self._update_ports_for_instance( [ 698.018672] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.018672] env[62235]: with excutils.save_and_reraise_exception(): [ 698.018672] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.018672] env[62235]: self.force_reraise() [ 698.018672] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.018672] env[62235]: raise self.value [ 698.018672] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.018672] env[62235]: updated_port = self._update_port( [ 698.018672] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.018672] env[62235]: _ensure_no_port_binding_failure(port) [ 698.018672] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.018672] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 698.019428] env[62235]: nova.exception.PortBindingFailed: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. [ 698.019428] env[62235]: Removing descriptor: 22 [ 698.019428] env[62235]: ERROR nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Traceback (most recent call last): [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] yield resources [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self.driver.spawn(context, instance, image_meta, [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.019428] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] vm_ref = self.build_virtual_machine(instance, [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] for vif in network_info: [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return self._sync_wrapper(fn, *args, **kwargs) [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self.wait() [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self[:] = self._gt.wait() [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return self._exit_event.wait() [ 698.019728] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] result = hub.switch() [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return self.greenlet.switch() [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] result = function(*args, **kwargs) [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return func(*args, **kwargs) [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] raise e [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] nwinfo = self.network_api.allocate_for_instance( [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.020063] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] created_port_ids = self._update_ports_for_instance( [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] with excutils.save_and_reraise_exception(): [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self.force_reraise() [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] raise self.value [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] updated_port = self._update_port( [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] _ensure_no_port_binding_failure(port) [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.020389] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] raise exception.PortBindingFailed(port_id=port['id']) [ 698.020736] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] nova.exception.PortBindingFailed: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. [ 698.020736] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] [ 698.020736] env[62235]: INFO nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Terminating instance [ 698.020736] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.020847] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquired lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.020997] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 698.077779] env[62235]: DEBUG nova.compute.utils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.083199] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 698.083199] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 698.126100] env[62235]: DEBUG nova.policy [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a43562ef6353442d8f41b056e4e007de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e262ac1b3084b3293071cd02f5dab5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 698.551847] env[62235]: DEBUG nova.compute.manager [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Received event network-changed-a11d8a7d-c492-408a-95ac-49da87650c44 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.553214] env[62235]: DEBUG nova.compute.manager [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Refreshing instance network info cache due to event network-changed-a11d8a7d-c492-408a-95ac-49da87650c44. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 698.553214] env[62235]: DEBUG oslo_concurrency.lockutils [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] Acquiring lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.554513] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.581443] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.622340] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Successfully created port: 98987fa9-752f-42e8-826b-80ad07983628 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.686621] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.029587] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "d165f549-1759-41bf-86fd-b77793bd49c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.029828] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "d165f549-1759-41bf-86fd-b77793bd49c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.042628] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15916fd5-1643-4f7f-9db0-5c6ddae64837 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.050275] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ced7f3-3022-4485-95ad-916f05d1f7b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.081578] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a95ee1-1592-4e46-8ef3-2a855955d1f5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.092019] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0e82fd-6fa0-4fc4-b8d0-251eb05e8784 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.096562] env[62235]: INFO nova.virt.block_device [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Booting with volume 3013693c-7ea8-46e4-b881-480284e2f3f6 at /dev/sda [ 699.109364] env[62235]: DEBUG nova.compute.provider_tree [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.165908] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-691229d8-e891-4f20-9932-045ef0664604 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.173584] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb89ee5-1f80-4315-8b76-66dfe6c014f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.195970] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Releasing lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.196371] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 699.196555] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.196959] env[62235]: DEBUG oslo_concurrency.lockutils [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] Acquired lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.197203] env[62235]: DEBUG nova.network.neutron [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Refreshing network info cache for port a11d8a7d-c492-408a-95ac-49da87650c44 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 699.198163] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28a9eeee-8821-4913-ba85-acc078c6f631 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.199804] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd63d256-5b58-4d57-80ba-b4b14cd79ded {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.209326] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82002baf-eee4-4f5a-a256-f64444787c65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.221931] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4087d905-daa3-4842-ac03-96a9bf254205 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.243299] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f245678-41f4-4685-87fc-822bf877d43e could not be found. [ 699.243523] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 699.243701] env[62235]: INFO nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 699.243939] env[62235]: DEBUG oslo.service.loopingcall [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.247496] env[62235]: DEBUG nova.compute.manager [-] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.247598] env[62235]: DEBUG nova.network.neutron [-] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 699.249954] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65725983-a531-4c70-8cda-1d3ac73fc00b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.256395] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb5e69e-8199-4c27-bd83-a1d5902c8791 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.269484] env[62235]: DEBUG nova.virt.block_device [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Updating existing volume attachment record: 73f94323-98f8-4a44-a535-afaf225c27c9 {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 699.273856] env[62235]: DEBUG nova.network.neutron [-] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.624723] env[62235]: DEBUG nova.scheduler.client.report [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.734413] env[62235]: DEBUG nova.network.neutron [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.779390] env[62235]: DEBUG nova.network.neutron [-] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.844157] env[62235]: DEBUG nova.network.neutron [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.070779] env[62235]: ERROR nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. [ 700.070779] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 700.070779] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.070779] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 700.070779] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.070779] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 700.070779] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.070779] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 700.070779] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.070779] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 700.070779] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.070779] env[62235]: ERROR nova.compute.manager raise self.value [ 700.070779] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.070779] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 700.070779] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.070779] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 700.071315] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.071315] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 700.071315] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. [ 700.071315] env[62235]: ERROR nova.compute.manager [ 700.071315] env[62235]: Traceback (most recent call last): [ 700.071315] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 700.071315] env[62235]: listener.cb(fileno) [ 700.071315] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.071315] env[62235]: result = function(*args, **kwargs) [ 700.071315] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 700.071315] env[62235]: return func(*args, **kwargs) [ 700.071315] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.071315] env[62235]: raise e [ 700.071315] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.071315] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 700.071315] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.071315] env[62235]: created_port_ids = self._update_ports_for_instance( [ 700.071315] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.071315] env[62235]: with excutils.save_and_reraise_exception(): [ 700.071315] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.071315] env[62235]: self.force_reraise() [ 700.071315] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.071315] env[62235]: raise self.value [ 700.071315] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.071315] env[62235]: updated_port = self._update_port( [ 700.071315] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.071315] env[62235]: _ensure_no_port_binding_failure(port) [ 700.071315] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.071315] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 700.072206] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. [ 700.072206] env[62235]: Removing descriptor: 22 [ 700.132805] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.133430] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.136661] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.342s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.138511] env[62235]: INFO nova.compute.claims [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.281887] env[62235]: INFO nova.compute.manager [-] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Took 1.03 seconds to deallocate network for instance. [ 700.286018] env[62235]: DEBUG nova.compute.claims [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 700.286227] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.347810] env[62235]: DEBUG oslo_concurrency.lockutils [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] Releasing lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.348209] env[62235]: DEBUG nova.compute.manager [req-7197d7d7-7891-4d6a-9167-99ac3135fa48 req-9b4816a4-7653-4559-8450-a6695cb1190e service nova] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Received event network-vif-deleted-a11d8a7d-c492-408a-95ac-49da87650c44 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.577208] env[62235]: DEBUG nova.compute.manager [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Received event network-changed-98987fa9-752f-42e8-826b-80ad07983628 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.577208] env[62235]: DEBUG nova.compute.manager [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Refreshing instance network info cache due to event network-changed-98987fa9-752f-42e8-826b-80ad07983628. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 700.577208] env[62235]: DEBUG oslo_concurrency.lockutils [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] Acquiring lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.577208] env[62235]: DEBUG oslo_concurrency.lockutils [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] Acquired lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.577208] env[62235]: DEBUG nova.network.neutron [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Refreshing network info cache for port 98987fa9-752f-42e8-826b-80ad07983628 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 700.638495] env[62235]: DEBUG nova.compute.utils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.641224] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.641425] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 700.691343] env[62235]: DEBUG nova.policy [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfee8aa24029459f871f9103a56bb70b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f27d43cbad4547808e1724b4ee2609ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.010375] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Successfully created port: c12de2f6-0783-42e7-a6f1-7c3e08dcae18 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.098954] env[62235]: DEBUG nova.network.neutron [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.141671] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 701.266431] env[62235]: DEBUG nova.network.neutron [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.389554] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.389554] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.389554] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.389554] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.389910] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.389910] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.389910] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.389910] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.389910] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.390068] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.390068] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.390068] env[62235]: DEBUG nova.virt.hardware [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.390482] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c84a23e-bd1d-4ac6-a5f4-22143f858906 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.398489] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe585ed-1b2d-4620-b50f-eee6d3ce4be7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.416323] env[62235]: ERROR nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Traceback (most recent call last): [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] yield resources [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self.driver.spawn(context, instance, image_meta, [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] vm_ref = self.build_virtual_machine(instance, [ 701.416323] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] for vif in network_info: [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] return self._sync_wrapper(fn, *args, **kwargs) [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self.wait() [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self[:] = self._gt.wait() [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] return self._exit_event.wait() [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.416909] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] current.throw(*self._exc) [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] result = function(*args, **kwargs) [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] return func(*args, **kwargs) [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] raise e [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] nwinfo = self.network_api.allocate_for_instance( [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] created_port_ids = self._update_ports_for_instance( [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] with excutils.save_and_reraise_exception(): [ 701.417600] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self.force_reraise() [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] raise self.value [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] updated_port = self._update_port( [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] _ensure_no_port_binding_failure(port) [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] raise exception.PortBindingFailed(port_id=port['id']) [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] nova.exception.PortBindingFailed: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. [ 701.418247] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] [ 701.418247] env[62235]: INFO nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Terminating instance [ 701.418768] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Acquiring lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.593638] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b606a8eb-342c-40d5-bcab-480d11e5e3c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.601653] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08afe1a7-5d03-4ec1-acb0-3054929e3bc8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.634356] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5219b5f2-0b66-4a1e-80aa-fa5f2555e82e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.645196] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fb9cae-3d51-48ef-849c-6b65ad02744c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.667278] env[62235]: DEBUG nova.compute.provider_tree [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.772057] env[62235]: DEBUG oslo_concurrency.lockutils [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] Releasing lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.772057] env[62235]: DEBUG nova.compute.manager [req-46d1c507-ac33-4b1a-8085-34a2111bca7f req-c663c743-455f-4722-8b7d-acb6c47fc960 service nova] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Received event network-vif-deleted-98987fa9-752f-42e8-826b-80ad07983628 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.772057] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Acquired lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.772057] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.023044] env[62235]: ERROR nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. [ 702.023044] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 702.023044] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.023044] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 702.023044] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.023044] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 702.023044] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.023044] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 702.023044] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.023044] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 702.023044] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.023044] env[62235]: ERROR nova.compute.manager raise self.value [ 702.023044] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.023044] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 702.023044] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.023044] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 702.023587] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.023587] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 702.023587] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. [ 702.023587] env[62235]: ERROR nova.compute.manager [ 702.023587] env[62235]: Traceback (most recent call last): [ 702.023587] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 702.023587] env[62235]: listener.cb(fileno) [ 702.023587] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.023587] env[62235]: result = function(*args, **kwargs) [ 702.023587] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.023587] env[62235]: return func(*args, **kwargs) [ 702.023587] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.023587] env[62235]: raise e [ 702.023587] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.023587] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 702.023587] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.023587] env[62235]: created_port_ids = self._update_ports_for_instance( [ 702.023587] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.023587] env[62235]: with excutils.save_and_reraise_exception(): [ 702.023587] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.023587] env[62235]: self.force_reraise() [ 702.023587] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.023587] env[62235]: raise self.value [ 702.023587] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.023587] env[62235]: updated_port = self._update_port( [ 702.023587] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.023587] env[62235]: _ensure_no_port_binding_failure(port) [ 702.023587] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.023587] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 702.024341] env[62235]: nova.exception.PortBindingFailed: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. [ 702.024341] env[62235]: Removing descriptor: 22 [ 702.161880] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.170337] env[62235]: DEBUG nova.scheduler.client.report [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.185521] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.185764] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.185922] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.186121] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.186272] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.186419] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.186619] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.186773] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.186937] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.187126] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.187273] env[62235]: DEBUG nova.virt.hardware [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.188451] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b897bd-a5cd-44d3-842e-2853f0abe998 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.197436] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65261af-a2fd-4ec7-83ff-29c766e16c4f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.212060] env[62235]: ERROR nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Traceback (most recent call last): [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] yield resources [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self.driver.spawn(context, instance, image_meta, [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] vm_ref = self.build_virtual_machine(instance, [ 702.212060] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] for vif in network_info: [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] return self._sync_wrapper(fn, *args, **kwargs) [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self.wait() [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self[:] = self._gt.wait() [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] return self._exit_event.wait() [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.212467] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] current.throw(*self._exc) [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] result = function(*args, **kwargs) [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] return func(*args, **kwargs) [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] raise e [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] nwinfo = self.network_api.allocate_for_instance( [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] created_port_ids = self._update_ports_for_instance( [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] with excutils.save_and_reraise_exception(): [ 702.212859] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self.force_reraise() [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] raise self.value [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] updated_port = self._update_port( [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] _ensure_no_port_binding_failure(port) [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] raise exception.PortBindingFailed(port_id=port['id']) [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] nova.exception.PortBindingFailed: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. [ 702.213288] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] [ 702.213288] env[62235]: INFO nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Terminating instance [ 702.213789] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.213950] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquired lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.214143] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.289385] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.384463] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.602073] env[62235]: DEBUG nova.compute.manager [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Received event network-changed-c12de2f6-0783-42e7-a6f1-7c3e08dcae18 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.602342] env[62235]: DEBUG nova.compute.manager [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Refreshing instance network info cache due to event network-changed-c12de2f6-0783-42e7-a6f1-7c3e08dcae18. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.602482] env[62235]: DEBUG oslo_concurrency.lockutils [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] Acquiring lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.675229] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.675751] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 702.678192] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.960s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.732629] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.810821] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.886798] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Releasing lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.887410] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 702.887723] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6052ab4a-e6b5-417e-bf49-0c7842446c64 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.896294] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546006ac-4c58-4a54-9823-d47fc1931671 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.917577] env[62235]: WARNING nova.virt.vmwareapi.driver [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 19e40767-d9f5-4b78-80f0-7f0f24464d40 could not be found. [ 702.917789] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 702.918069] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59603060-3f14-420f-8862-3eecdaaeac6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.925422] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78db29c-50a9-4178-94cd-487fe93aa63f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.946010] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 19e40767-d9f5-4b78-80f0-7f0f24464d40 could not be found. [ 702.946328] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 702.946582] env[62235]: INFO nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Took 0.06 seconds to destroy the instance on the hypervisor. [ 702.946846] env[62235]: DEBUG oslo.service.loopingcall [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.947082] env[62235]: DEBUG nova.compute.manager [-] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.947178] env[62235]: DEBUG nova.network.neutron [-] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.967698] env[62235]: DEBUG nova.network.neutron [-] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.182255] env[62235]: DEBUG nova.compute.utils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.183681] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.183849] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 703.244205] env[62235]: DEBUG nova.policy [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfee8aa24029459f871f9103a56bb70b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f27d43cbad4547808e1724b4ee2609ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 703.313389] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Releasing lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.313812] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.314010] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.314329] env[62235]: DEBUG oslo_concurrency.lockutils [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] Acquired lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.314501] env[62235]: DEBUG nova.network.neutron [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Refreshing network info cache for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.315530] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae22dc86-cfcc-4596-bb92-db062a1b3d4f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.332093] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcf933d-5007-495a-bbe7-16b3738346ec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.356088] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b348bc9f-3bd4-475f-9957-42a99848005c could not be found. [ 703.357035] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.357411] env[62235]: INFO nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 703.357720] env[62235]: DEBUG oslo.service.loopingcall [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.360408] env[62235]: DEBUG nova.compute.manager [-] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.360511] env[62235]: DEBUG nova.network.neutron [-] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.386195] env[62235]: DEBUG nova.network.neutron [-] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.471813] env[62235]: DEBUG nova.network.neutron [-] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.567687] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Successfully created port: ccacb20d-8b17-4df3-afec-4e31be46daa7 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 703.589996] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31672b42-5364-4ec4-a732-10ccb24ebde9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.597562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b499e7-3cd5-4211-b0b7-c217902bf2d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.628980] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbdc994-fa2f-4db7-a94b-329d0084cf83 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.636351] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e296eb16-722a-4865-b74b-9dcd53cbec70 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.649852] env[62235]: DEBUG nova.compute.provider_tree [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.690062] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.846750] env[62235]: DEBUG nova.network.neutron [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.888346] env[62235]: DEBUG nova.network.neutron [-] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.976458] env[62235]: INFO nova.compute.manager [-] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Took 1.03 seconds to deallocate network for instance. [ 703.982428] env[62235]: DEBUG nova.network.neutron [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.155259] env[62235]: DEBUG nova.scheduler.client.report [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.391469] env[62235]: INFO nova.compute.manager [-] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Took 1.03 seconds to deallocate network for instance. [ 704.394289] env[62235]: DEBUG nova.compute.claims [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.394508] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.483554] env[62235]: DEBUG oslo_concurrency.lockutils [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] Releasing lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.484517] env[62235]: DEBUG nova.compute.manager [req-9d4d697b-d4ca-4da5-b19c-35b608179daf req-061e8fb1-d2eb-4921-a0e1-4d024782e3a4 service nova] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Received event network-vif-deleted-c12de2f6-0783-42e7-a6f1-7c3e08dcae18 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.545240] env[62235]: INFO nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Took 0.57 seconds to detach 1 volumes for instance. [ 704.548578] env[62235]: DEBUG nova.compute.claims [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.548733] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.564129] env[62235]: ERROR nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. [ 704.564129] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 704.564129] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.564129] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 704.564129] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.564129] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 704.564129] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.564129] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 704.564129] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.564129] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 704.564129] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.564129] env[62235]: ERROR nova.compute.manager raise self.value [ 704.564129] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.564129] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 704.564129] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.564129] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 704.564810] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.564810] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 704.564810] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. [ 704.564810] env[62235]: ERROR nova.compute.manager [ 704.564810] env[62235]: Traceback (most recent call last): [ 704.564810] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 704.564810] env[62235]: listener.cb(fileno) [ 704.564810] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.564810] env[62235]: result = function(*args, **kwargs) [ 704.564810] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.564810] env[62235]: return func(*args, **kwargs) [ 704.564810] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.564810] env[62235]: raise e [ 704.564810] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.564810] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 704.564810] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.564810] env[62235]: created_port_ids = self._update_ports_for_instance( [ 704.564810] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.564810] env[62235]: with excutils.save_and_reraise_exception(): [ 704.564810] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.564810] env[62235]: self.force_reraise() [ 704.564810] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.564810] env[62235]: raise self.value [ 704.564810] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.564810] env[62235]: updated_port = self._update_port( [ 704.564810] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.564810] env[62235]: _ensure_no_port_binding_failure(port) [ 704.564810] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.564810] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 704.565786] env[62235]: nova.exception.PortBindingFailed: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. [ 704.565786] env[62235]: Removing descriptor: 22 [ 704.625056] env[62235]: DEBUG nova.compute.manager [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Received event network-changed-ccacb20d-8b17-4df3-afec-4e31be46daa7 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.625262] env[62235]: DEBUG nova.compute.manager [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Refreshing instance network info cache due to event network-changed-ccacb20d-8b17-4df3-afec-4e31be46daa7. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 704.625478] env[62235]: DEBUG oslo_concurrency.lockutils [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] Acquiring lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.625623] env[62235]: DEBUG oslo_concurrency.lockutils [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] Acquired lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.625781] env[62235]: DEBUG nova.network.neutron [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Refreshing network info cache for port ccacb20d-8b17-4df3-afec-4e31be46daa7 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 704.661047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.661674] env[62235]: ERROR nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Traceback (most recent call last): [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self.driver.spawn(context, instance, image_meta, [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] vm_ref = self.build_virtual_machine(instance, [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.661674] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] for vif in network_info: [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return self._sync_wrapper(fn, *args, **kwargs) [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self.wait() [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self[:] = self._gt.wait() [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return self._exit_event.wait() [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] result = hub.switch() [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.662040] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return self.greenlet.switch() [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] result = function(*args, **kwargs) [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] return func(*args, **kwargs) [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] raise e [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] nwinfo = self.network_api.allocate_for_instance( [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] created_port_ids = self._update_ports_for_instance( [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] with excutils.save_and_reraise_exception(): [ 704.662459] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] self.force_reraise() [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] raise self.value [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] updated_port = self._update_port( [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] _ensure_no_port_binding_failure(port) [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] raise exception.PortBindingFailed(port_id=port['id']) [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] nova.exception.PortBindingFailed: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. [ 704.662847] env[62235]: ERROR nova.compute.manager [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] [ 704.663200] env[62235]: DEBUG nova.compute.utils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 704.663641] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.885s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.667138] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Build of instance 584b41c1-976d-49e9-b5ea-c5c22e1877c5 was re-scheduled: Binding failed for port c5eb8ae5-46e9-4185-8194-9ee04583ef3f, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 704.667637] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 704.667719] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Acquiring lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.667804] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Acquired lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.667954] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.702923] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.731899] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.732198] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.732366] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.732545] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.732685] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.732827] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.733069] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.733234] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.733401] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.733586] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.733766] env[62235]: DEBUG nova.virt.hardware [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.734838] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a312e6f4-04da-463c-91c6-878809965898 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.742962] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f907d033-735b-4ea8-8799-ad23939d7804 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.756290] env[62235]: ERROR nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Traceback (most recent call last): [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] yield resources [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self.driver.spawn(context, instance, image_meta, [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] vm_ref = self.build_virtual_machine(instance, [ 704.756290] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] for vif in network_info: [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] return self._sync_wrapper(fn, *args, **kwargs) [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self.wait() [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self[:] = self._gt.wait() [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] return self._exit_event.wait() [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 704.756721] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] current.throw(*self._exc) [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] result = function(*args, **kwargs) [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] return func(*args, **kwargs) [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] raise e [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] nwinfo = self.network_api.allocate_for_instance( [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] created_port_ids = self._update_ports_for_instance( [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] with excutils.save_and_reraise_exception(): [ 704.757127] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self.force_reraise() [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] raise self.value [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] updated_port = self._update_port( [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] _ensure_no_port_binding_failure(port) [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] raise exception.PortBindingFailed(port_id=port['id']) [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] nova.exception.PortBindingFailed: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. [ 704.757557] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] [ 704.757557] env[62235]: INFO nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Terminating instance [ 704.758703] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.145028] env[62235]: DEBUG nova.network.neutron [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.199473] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.295866] env[62235]: DEBUG nova.network.neutron [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.329952] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.525732] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4baf77e-943f-4b00-a28a-e9cadac9b826 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.533302] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2674a448-b984-4b43-920f-71b6fdf00059 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.562342] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76ff369-70bf-4be1-9bed-fd9f0164c7ed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.569693] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dbb3f7-cc82-4060-aa61-c8ed3384ffbe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.584145] env[62235]: DEBUG nova.compute.provider_tree [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.799072] env[62235]: DEBUG oslo_concurrency.lockutils [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] Releasing lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.799072] env[62235]: DEBUG nova.compute.manager [req-ab7eb328-5d6e-4185-b471-ababe7fd2585 req-aa7b033d-9a01-4f6f-a777-538350bc4467 service nova] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Received event network-vif-deleted-ccacb20d-8b17-4df3-afec-4e31be46daa7 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.799220] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquired lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.799381] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.835473] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Releasing lock "refresh_cache-584b41c1-976d-49e9-b5ea-c5c22e1877c5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.835710] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 705.835888] env[62235]: DEBUG nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.836060] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.851239] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.087277] env[62235]: DEBUG nova.scheduler.client.report [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.325007] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.354367] env[62235]: DEBUG nova.network.neutron [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.408850] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.593627] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.593627] env[62235]: ERROR nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. [ 706.593627] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Traceback (most recent call last): [ 706.593627] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.593627] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self.driver.spawn(context, instance, image_meta, [ 706.593627] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.593627] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.593627] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.593627] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] vm_ref = self.build_virtual_machine(instance, [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] for vif in network_info: [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return self._sync_wrapper(fn, *args, **kwargs) [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self.wait() [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self[:] = self._gt.wait() [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return self._exit_event.wait() [ 706.594221] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] result = hub.switch() [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return self.greenlet.switch() [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] result = function(*args, **kwargs) [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] return func(*args, **kwargs) [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] raise e [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] nwinfo = self.network_api.allocate_for_instance( [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.594582] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] created_port_ids = self._update_ports_for_instance( [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] with excutils.save_and_reraise_exception(): [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] self.force_reraise() [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] raise self.value [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] updated_port = self._update_port( [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] _ensure_no_port_binding_failure(port) [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.594923] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] raise exception.PortBindingFailed(port_id=port['id']) [ 706.595288] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] nova.exception.PortBindingFailed: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. [ 706.595288] env[62235]: ERROR nova.compute.manager [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] [ 706.595288] env[62235]: DEBUG nova.compute.utils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.595407] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.710s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.597064] env[62235]: INFO nova.compute.claims [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.599627] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Build of instance 4d10be8a-a4ba-46f7-b4da-c37f8520bab6 was re-scheduled: Binding failed for port c3347862-c63a-4a4f-9f97-ca812573ba26, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.600230] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.600456] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.600600] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.600752] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.860395] env[62235]: INFO nova.compute.manager [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] [instance: 584b41c1-976d-49e9-b5ea-c5c22e1877c5] Took 1.02 seconds to deallocate network for instance. [ 706.911276] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Releasing lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.911709] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.911911] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.912298] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1449393-fabb-4e97-996a-b32ec0688a1c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.921142] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbe3bca-79ee-44a2-b869-cd73f9c18934 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.942537] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 012dd63b-a092-4a1d-a79c-1e1bf2fd8771 could not be found. [ 706.942745] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.942925] env[62235]: INFO nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Took 0.03 seconds to destroy the instance on the hypervisor. [ 706.943199] env[62235]: DEBUG oslo.service.loopingcall [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.943917] env[62235]: DEBUG nova.compute.manager [-] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.943917] env[62235]: DEBUG nova.network.neutron [-] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.961460] env[62235]: DEBUG nova.network.neutron [-] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.125381] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.210814] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.464224] env[62235]: DEBUG nova.network.neutron [-] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.721030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-4d10be8a-a4ba-46f7-b4da-c37f8520bab6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.721030] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 707.721030] env[62235]: DEBUG nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.721030] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.738295] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.886030] env[62235]: INFO nova.scheduler.client.report [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Deleted allocations for instance 584b41c1-976d-49e9-b5ea-c5c22e1877c5 [ 707.948602] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d3ab4e-8772-48c3-a08b-152784bc2eb9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.956659] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc70756-000b-48c1-837b-98412f61fed0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.985399] env[62235]: INFO nova.compute.manager [-] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Took 1.04 seconds to deallocate network for instance. [ 707.987976] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bea028-8bb4-4de2-ab29-c6ed7ec663f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.990591] env[62235]: DEBUG nova.compute.claims [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 707.990768] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.995508] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961b7289-b5f7-414f-bd27-e046d4da7a9f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.008396] env[62235]: DEBUG nova.compute.provider_tree [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.241975] env[62235]: DEBUG nova.network.neutron [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.397024] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e1d870c7-c209-402f-a294-01a2502b4b48 tempest-AttachInterfacesV270Test-43710041 tempest-AttachInterfacesV270Test-43710041-project-member] Lock "584b41c1-976d-49e9-b5ea-c5c22e1877c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.745s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.511107] env[62235]: DEBUG nova.scheduler.client.report [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.744593] env[62235]: INFO nova.compute.manager [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 4d10be8a-a4ba-46f7-b4da-c37f8520bab6] Took 1.02 seconds to deallocate network for instance. [ 708.899031] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.016167] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.016167] env[62235]: DEBUG nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.023979] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.210s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.024898] env[62235]: INFO nova.compute.claims [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.433706] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.534310] env[62235]: DEBUG nova.compute.utils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.537696] env[62235]: DEBUG nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Not allocating networking since 'none' was specified. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 709.783577] env[62235]: INFO nova.scheduler.client.report [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleted allocations for instance 4d10be8a-a4ba-46f7-b4da-c37f8520bab6 [ 710.041416] env[62235]: DEBUG nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.298031] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c5cdf262-57bb-4c21-b30f-cfc69060e432 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "4d10be8a-a4ba-46f7-b4da-c37f8520bab6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.703s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.441118] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad64fcd-cabf-4896-9b29-5825109c2d9b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.447115] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3702b3f4-c2bb-4e3a-8b35-4e2165d912c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.479730] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5ecc1a-eb08-439a-b3c5-1eb1b6eae65f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.487450] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787d0207-6494-489f-8d5f-fa327f001e12 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.508637] env[62235]: DEBUG nova.compute.provider_tree [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.801839] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.011881] env[62235]: DEBUG nova.scheduler.client.report [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.052650] env[62235]: DEBUG nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.079597] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.079839] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.079996] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.080234] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.080396] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.080546] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.080753] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.080927] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.081113] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.081273] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.081456] env[62235]: DEBUG nova.virt.hardware [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.082392] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f4a43d-c598-42ba-9102-2bf20356797b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.090270] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00138f6-b9f9-4924-bfcb-fec8fb1af2c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.105906] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 711.111456] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Creating folder: Project (c1a8934577264862b1edd8a24ca35214). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 711.111732] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bfbeca8-28c4-4eeb-88e8-fa8ce3a7355b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.124391] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Created folder: Project (c1a8934577264862b1edd8a24ca35214) in parent group-v273362. [ 711.124580] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Creating folder: Instances. Parent ref: group-v273379. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 711.124804] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70b129a4-fbec-4c72-9a5f-2bc7772f57e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.133781] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Created folder: Instances in parent group-v273379. [ 711.134025] env[62235]: DEBUG oslo.service.loopingcall [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.134223] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 711.134414] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc6163ac-e292-4f52-ac4c-acb9b29e54a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.149788] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 711.149788] env[62235]: value = "task-1271472" [ 711.149788] env[62235]: _type = "Task" [ 711.149788] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.157757] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271472, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.324317] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.519017] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.519017] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 711.520728] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.779s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.522935] env[62235]: INFO nova.compute.claims [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.660057] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271472, 'name': CreateVM_Task, 'duration_secs': 0.263317} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.660239] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 711.660670] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.660867] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.661159] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 711.661400] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4917ee4-74b8-45d8-8dbb-34e5ff6fe6a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.665904] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 711.665904] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a30dd-d6fa-9b77-3328-443052708ef2" [ 711.665904] env[62235]: _type = "Task" [ 711.665904] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.673830] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a30dd-d6fa-9b77-3328-443052708ef2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.022810] env[62235]: DEBUG nova.compute.utils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.024276] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.024396] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.093201] env[62235]: DEBUG nova.policy [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '047e0cee73984f6396b9732cc1d47125', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07d29daa0c73431eb3b58b9f64ecb94a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.176929] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a30dd-d6fa-9b77-3328-443052708ef2, 'name': SearchDatastore_Task, 'duration_secs': 0.010886} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.177260] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.177493] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.177722] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.177919] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.178042] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.178306] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7d8cd30-43e0-4d1b-9080-a0ea8dea8842 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.187313] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.187493] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 712.188201] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a74465bb-d965-4e38-bec2-d241ffa6c84b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.193711] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 712.193711] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52088215-a5d1-cb16-597c-82056eac0514" [ 712.193711] env[62235]: _type = "Task" [ 712.193711] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.202731] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52088215-a5d1-cb16-597c-82056eac0514, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.529375] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 712.563213] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Successfully created port: 580a74e8-7dce-4bdd-b04f-aa1f1f68e503 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.707876] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52088215-a5d1-cb16-597c-82056eac0514, 'name': SearchDatastore_Task, 'duration_secs': 0.00796} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.710950] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a866f9ad-dffd-407f-9ca0-1443a0472fd3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.720365] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 712.720365] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e4fe8f-bc97-83d9-459d-03b9e4742a65" [ 712.720365] env[62235]: _type = "Task" [ 712.720365] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.728959] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e4fe8f-bc97-83d9-459d-03b9e4742a65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.953771] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e983448-9fb6-436b-b4e6-17535f7d828a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.962114] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c014459-1ab7-4972-9409-47785378594e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.994781] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271ecb38-c891-4208-9e11-8428cfc0a9d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.003336] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28e59f6-d396-4842-90d8-0987a1a79132 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.024465] env[62235]: DEBUG nova.compute.provider_tree [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.232470] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e4fe8f-bc97-83d9-459d-03b9e4742a65, 'name': SearchDatastore_Task, 'duration_secs': 0.009571} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.232470] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.232470] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] be4fb088-8b01-4245-82fb-b0e085825597/be4fb088-8b01-4245-82fb-b0e085825597.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 713.232470] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30a2efd9-317f-4a10-b3ff-91a439abff97 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.243297] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 713.243297] env[62235]: value = "task-1271473" [ 713.243297] env[62235]: _type = "Task" [ 713.243297] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.250250] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271473, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.505285] env[62235]: DEBUG nova.compute.manager [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Received event network-changed-580a74e8-7dce-4bdd-b04f-aa1f1f68e503 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.505460] env[62235]: DEBUG nova.compute.manager [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Refreshing instance network info cache due to event network-changed-580a74e8-7dce-4bdd-b04f-aa1f1f68e503. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 713.505713] env[62235]: DEBUG oslo_concurrency.lockutils [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] Acquiring lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.505809] env[62235]: DEBUG oslo_concurrency.lockutils [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] Acquired lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.505960] env[62235]: DEBUG nova.network.neutron [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Refreshing network info cache for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.527558] env[62235]: DEBUG nova.scheduler.client.report [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.540489] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 713.574420] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 713.574672] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 713.574835] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.574999] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 713.576830] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.577038] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 713.581653] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 713.581653] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 713.581653] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 713.581653] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 713.581653] env[62235]: DEBUG nova.virt.hardware [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.582869] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4843f53-6187-4c4d-929e-610f4e2d8a7c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.592142] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aab1531-917d-467d-89c4-4300465f9e97 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.674999] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.675904] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "ced2aa67-ce75-4af5-9c55-773835d05733" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.752694] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271473, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442812} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.752885] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] be4fb088-8b01-4245-82fb-b0e085825597/be4fb088-8b01-4245-82fb-b0e085825597.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 713.753067] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 713.753330] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5527a595-f830-431d-8590-d2cb656d2907 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.759259] env[62235]: ERROR nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. [ 713.759259] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.759259] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.759259] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.759259] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.759259] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.759259] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.759259] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.759259] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.759259] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 713.759259] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.759259] env[62235]: ERROR nova.compute.manager raise self.value [ 713.759259] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.759259] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.759259] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.759259] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.759737] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.759737] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.759737] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. [ 713.759737] env[62235]: ERROR nova.compute.manager [ 713.759737] env[62235]: Traceback (most recent call last): [ 713.759737] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.759737] env[62235]: listener.cb(fileno) [ 713.759737] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.759737] env[62235]: result = function(*args, **kwargs) [ 713.759737] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.759737] env[62235]: return func(*args, **kwargs) [ 713.759737] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.759737] env[62235]: raise e [ 713.759737] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.759737] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 713.759737] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.759737] env[62235]: created_port_ids = self._update_ports_for_instance( [ 713.759737] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.759737] env[62235]: with excutils.save_and_reraise_exception(): [ 713.759737] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.759737] env[62235]: self.force_reraise() [ 713.759737] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.759737] env[62235]: raise self.value [ 713.759737] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.759737] env[62235]: updated_port = self._update_port( [ 713.759737] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.759737] env[62235]: _ensure_no_port_binding_failure(port) [ 713.759737] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.759737] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.760543] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. [ 713.760543] env[62235]: Removing descriptor: 16 [ 713.760543] env[62235]: ERROR nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Traceback (most recent call last): [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] yield resources [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self.driver.spawn(context, instance, image_meta, [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.760543] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] vm_ref = self.build_virtual_machine(instance, [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] for vif in network_info: [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return self._sync_wrapper(fn, *args, **kwargs) [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self.wait() [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self[:] = self._gt.wait() [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return self._exit_event.wait() [ 713.760933] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] result = hub.switch() [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return self.greenlet.switch() [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] result = function(*args, **kwargs) [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return func(*args, **kwargs) [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] raise e [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] nwinfo = self.network_api.allocate_for_instance( [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.761299] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] created_port_ids = self._update_ports_for_instance( [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] with excutils.save_and_reraise_exception(): [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self.force_reraise() [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] raise self.value [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] updated_port = self._update_port( [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] _ensure_no_port_binding_failure(port) [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.761656] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] raise exception.PortBindingFailed(port_id=port['id']) [ 713.761975] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] nova.exception.PortBindingFailed: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. [ 713.761975] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] [ 713.761975] env[62235]: INFO nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Terminating instance [ 713.762908] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 713.762908] env[62235]: value = "task-1271474" [ 713.762908] env[62235]: _type = "Task" [ 713.762908] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.763784] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Acquiring lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.773466] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.021893] env[62235]: DEBUG nova.network.neutron [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.032829] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.033109] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.036038] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.179s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.103504] env[62235]: DEBUG nova.network.neutron [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.273464] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066841} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.273798] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.274625] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e6f66a-b712-4cd4-9895-dbcacb7ba0f5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.294815] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] be4fb088-8b01-4245-82fb-b0e085825597/be4fb088-8b01-4245-82fb-b0e085825597.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.294815] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-095fe3b1-b559-4398-b80b-2966f6022296 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.315097] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 714.315097] env[62235]: value = "task-1271475" [ 714.315097] env[62235]: _type = "Task" [ 714.315097] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.322944] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271475, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.540363] env[62235]: DEBUG nova.compute.utils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.544654] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 714.544892] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 714.583312] env[62235]: DEBUG nova.policy [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eff208ffe4814720bb6cc29dd1cdb6c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1f2551ad3cfb4a0d8112f85d4e95fe64', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 714.606929] env[62235]: DEBUG oslo_concurrency.lockutils [req-6f17ec9a-0ae2-42ce-b127-c2306937ea3d req-5651808f-4f63-4513-86af-027ad4edfbfb service nova] Releasing lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.606983] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Acquired lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.607204] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.828179] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.891914] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eadac4d-ffe5-4578-8ef2-29d43c582076 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.900639] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2158afd8-f547-466b-80da-df59d7f349b4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.932100] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02bac20-283b-4beb-84e2-7bf93b54bb42 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.935554] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Successfully created port: c0733950-c935-475d-9191-85daffca842b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.942015] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724857a8-8571-42c6-9bc7-8f4d31f7e5a1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.956311] env[62235]: DEBUG nova.compute.provider_tree [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.048425] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.133185] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.274141] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.326895] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271475, 'name': ReconfigVM_Task, 'duration_secs': 0.857483} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.327488] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Reconfigured VM instance instance-00000028 to attach disk [datastore2] be4fb088-8b01-4245-82fb-b0e085825597/be4fb088-8b01-4245-82fb-b0e085825597.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.328205] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fab5b9b9-8c33-48ba-9ef5-4ccfbd2297f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.334303] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 715.334303] env[62235]: value = "task-1271476" [ 715.334303] env[62235]: _type = "Task" [ 715.334303] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.341878] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271476, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.463292] env[62235]: DEBUG nova.scheduler.client.report [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.531970] env[62235]: DEBUG nova.compute.manager [req-e79d28fd-f510-4bc5-b659-ad398dd94543 req-a1531a65-4002-4d65-ab16-8c37f5b91a5d service nova] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Received event network-vif-deleted-580a74e8-7dce-4bdd-b04f-aa1f1f68e503 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.776956] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Releasing lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.777465] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.777592] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.777891] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d48e3c3-d359-4c96-a9f2-f8d443db3e30 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.786859] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd4f792-1234-430c-901f-ade5365250c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.811020] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 693db195-ce52-40ff-bbb4-91f03d5a0bd9 could not be found. [ 715.811020] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.811020] env[62235]: INFO nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 715.811020] env[62235]: DEBUG oslo.service.loopingcall [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.811020] env[62235]: DEBUG nova.compute.manager [-] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.811020] env[62235]: DEBUG nova.network.neutron [-] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.842843] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271476, 'name': Rename_Task, 'duration_secs': 0.122315} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.843999] env[62235]: DEBUG nova.network.neutron [-] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.845372] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 715.847728] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-617b9f1b-41a4-401b-9bc3-ea6949b2f9a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.854084] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 715.854084] env[62235]: value = "task-1271477" [ 715.854084] env[62235]: _type = "Task" [ 715.854084] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.865528] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.971313] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.971947] env[62235]: ERROR nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Traceback (most recent call last): [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self.driver.spawn(context, instance, image_meta, [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] vm_ref = self.build_virtual_machine(instance, [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.971947] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] for vif in network_info: [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] return self._sync_wrapper(fn, *args, **kwargs) [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self.wait() [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self[:] = self._gt.wait() [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] return self._exit_event.wait() [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] current.throw(*self._exc) [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.972357] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] result = function(*args, **kwargs) [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] return func(*args, **kwargs) [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] raise e [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] nwinfo = self.network_api.allocate_for_instance( [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] created_port_ids = self._update_ports_for_instance( [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] with excutils.save_and_reraise_exception(): [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] self.force_reraise() [ 715.972781] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] raise self.value [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] updated_port = self._update_port( [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] _ensure_no_port_binding_failure(port) [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] raise exception.PortBindingFailed(port_id=port['id']) [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] nova.exception.PortBindingFailed: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. [ 715.973203] env[62235]: ERROR nova.compute.manager [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] [ 715.973203] env[62235]: DEBUG nova.compute.utils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 715.973952] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.688s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.977052] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Build of instance 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2 was re-scheduled: Binding failed for port 35b9ca76-9b0e-4441-8212-f56017629258, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 715.977204] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 715.977347] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Acquiring lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.977497] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Acquired lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.977684] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 716.059126] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.090862] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.092225] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.092420] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.092860] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.093104] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.093316] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.093585] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.094155] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.094500] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.096512] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.096718] env[62235]: DEBUG nova.virt.hardware [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.097600] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49feaf8-d54b-4156-8542-c22c8af9225a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.107015] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3543378-d6be-4879-81c8-f6a8ada4661e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.177469] env[62235]: ERROR nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. [ 716.177469] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.177469] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.177469] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.177469] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.177469] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.177469] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.177469] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.177469] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.177469] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 716.177469] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.177469] env[62235]: ERROR nova.compute.manager raise self.value [ 716.177469] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.177469] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.177469] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.177469] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.178078] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.178078] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.178078] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. [ 716.178078] env[62235]: ERROR nova.compute.manager [ 716.178078] env[62235]: Traceback (most recent call last): [ 716.178078] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.178078] env[62235]: listener.cb(fileno) [ 716.178078] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.178078] env[62235]: result = function(*args, **kwargs) [ 716.178078] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.178078] env[62235]: return func(*args, **kwargs) [ 716.178078] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.178078] env[62235]: raise e [ 716.178078] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.178078] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 716.178078] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.178078] env[62235]: created_port_ids = self._update_ports_for_instance( [ 716.178078] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.178078] env[62235]: with excutils.save_and_reraise_exception(): [ 716.178078] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.178078] env[62235]: self.force_reraise() [ 716.178078] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.178078] env[62235]: raise self.value [ 716.178078] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.178078] env[62235]: updated_port = self._update_port( [ 716.178078] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.178078] env[62235]: _ensure_no_port_binding_failure(port) [ 716.178078] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.178078] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.178894] env[62235]: nova.exception.PortBindingFailed: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. [ 716.178894] env[62235]: Removing descriptor: 16 [ 716.178894] env[62235]: ERROR nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Traceback (most recent call last): [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] yield resources [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self.driver.spawn(context, instance, image_meta, [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.178894] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] vm_ref = self.build_virtual_machine(instance, [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] for vif in network_info: [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return self._sync_wrapper(fn, *args, **kwargs) [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self.wait() [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self[:] = self._gt.wait() [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return self._exit_event.wait() [ 716.179285] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] result = hub.switch() [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return self.greenlet.switch() [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] result = function(*args, **kwargs) [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return func(*args, **kwargs) [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] raise e [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] nwinfo = self.network_api.allocate_for_instance( [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.179735] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] created_port_ids = self._update_ports_for_instance( [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] with excutils.save_and_reraise_exception(): [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self.force_reraise() [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] raise self.value [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] updated_port = self._update_port( [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] _ensure_no_port_binding_failure(port) [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.180186] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] raise exception.PortBindingFailed(port_id=port['id']) [ 716.180602] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] nova.exception.PortBindingFailed: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. [ 716.180602] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] [ 716.180602] env[62235]: INFO nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Terminating instance [ 716.181484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.181650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquired lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.181816] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 716.348496] env[62235]: DEBUG nova.network.neutron [-] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.364763] env[62235]: DEBUG oslo_vmware.api [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271477, 'name': PowerOnVM_Task, 'duration_secs': 0.404585} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.365051] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 716.365252] env[62235]: INFO nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Took 5.31 seconds to spawn the instance on the hypervisor. [ 716.365427] env[62235]: DEBUG nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 716.366190] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ffe86d-23f9-4b49-9337-3424891379e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.498290] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.589615] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.697225] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.777766] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.804715] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5622eaeb-d34e-459b-af5c-71503f4e39a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.812707] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89ce8ca-7c26-4cad-b06e-167abf12ee25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.844278] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1a58bb-aa5a-43b2-a28c-857b595039fa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.851272] env[62235]: INFO nova.compute.manager [-] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Took 1.04 seconds to deallocate network for instance. [ 716.854071] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cdac06-522b-43e3-909f-2b89269541c6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.858357] env[62235]: DEBUG nova.compute.claims [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 716.858537] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.868342] env[62235]: DEBUG nova.compute.provider_tree [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.880832] env[62235]: INFO nova.compute.manager [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Took 41.03 seconds to build instance. [ 717.092596] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Releasing lock "refresh_cache-2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.092862] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 717.092938] env[62235]: DEBUG nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.093122] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.107482] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.280510] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Releasing lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.280934] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 717.281140] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 717.281436] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4f38a1d-0f3b-4e30-a18c-8d086e78241e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.290861] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbb9050-1903-4347-bfc2-de7a344f862e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.314341] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 452cf04e-d6e9-48a8-a77b-f931c9440858 could not be found. [ 717.314552] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 717.314797] env[62235]: INFO nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Took 0.03 seconds to destroy the instance on the hypervisor. [ 717.315055] env[62235]: DEBUG oslo.service.loopingcall [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.315325] env[62235]: DEBUG nova.compute.manager [-] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.315422] env[62235]: DEBUG nova.network.neutron [-] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.334343] env[62235]: DEBUG nova.network.neutron [-] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.371379] env[62235]: DEBUG nova.scheduler.client.report [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.382840] env[62235]: DEBUG oslo_concurrency.lockutils [None req-064e80a9-e421-4d0b-8f19-1e07aa73f9a3 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "be4fb088-8b01-4245-82fb-b0e085825597" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.519s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.556800] env[62235]: DEBUG nova.compute.manager [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Received event network-changed-c0733950-c935-475d-9191-85daffca842b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.557035] env[62235]: DEBUG nova.compute.manager [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Refreshing instance network info cache due to event network-changed-c0733950-c935-475d-9191-85daffca842b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 717.557212] env[62235]: DEBUG oslo_concurrency.lockutils [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] Acquiring lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.557354] env[62235]: DEBUG oslo_concurrency.lockutils [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] Acquired lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.557516] env[62235]: DEBUG nova.network.neutron [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Refreshing network info cache for port c0733950-c935-475d-9191-85daffca842b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.609977] env[62235]: DEBUG nova.network.neutron [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.706855] env[62235]: DEBUG nova.compute.manager [None req-d6c50e2f-61f0-43c3-9b98-053d6a7f33dd tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 717.707700] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da3f3ab-3389-4948-9ea5-8f0eb67b7385 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.766636] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "be4fb088-8b01-4245-82fb-b0e085825597" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.766945] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "be4fb088-8b01-4245-82fb-b0e085825597" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.767209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "be4fb088-8b01-4245-82fb-b0e085825597-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.767426] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "be4fb088-8b01-4245-82fb-b0e085825597-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.767626] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "be4fb088-8b01-4245-82fb-b0e085825597-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.769585] env[62235]: INFO nova.compute.manager [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Terminating instance [ 717.771258] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "refresh_cache-be4fb088-8b01-4245-82fb-b0e085825597" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.771408] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquired lock "refresh_cache-be4fb088-8b01-4245-82fb-b0e085825597" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.771571] env[62235]: DEBUG nova.network.neutron [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.837601] env[62235]: DEBUG nova.network.neutron [-] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.877645] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.878314] env[62235]: ERROR nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Traceback (most recent call last): [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self.driver.spawn(context, instance, image_meta, [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] vm_ref = self.build_virtual_machine(instance, [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.878314] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] for vif in network_info: [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return self._sync_wrapper(fn, *args, **kwargs) [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self.wait() [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self[:] = self._gt.wait() [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return self._exit_event.wait() [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] result = hub.switch() [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 717.878678] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return self.greenlet.switch() [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] result = function(*args, **kwargs) [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] return func(*args, **kwargs) [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] raise e [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] nwinfo = self.network_api.allocate_for_instance( [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] created_port_ids = self._update_ports_for_instance( [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] with excutils.save_and_reraise_exception(): [ 717.879084] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] self.force_reraise() [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] raise self.value [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] updated_port = self._update_port( [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] _ensure_no_port_binding_failure(port) [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] raise exception.PortBindingFailed(port_id=port['id']) [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] nova.exception.PortBindingFailed: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. [ 717.879467] env[62235]: ERROR nova.compute.manager [instance: 3f245678-41f4-4685-87fc-822bf877d43e] [ 717.879803] env[62235]: DEBUG nova.compute.utils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.880200] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.486s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.883212] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Build of instance 3f245678-41f4-4685-87fc-822bf877d43e was re-scheduled: Binding failed for port a11d8a7d-c492-408a-95ac-49da87650c44, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.883638] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.883862] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquiring lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.884013] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Acquired lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.884188] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.885479] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.081218] env[62235]: DEBUG nova.network.neutron [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.113735] env[62235]: INFO nova.compute.manager [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] [instance: 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2] Took 1.02 seconds to deallocate network for instance. [ 718.159786] env[62235]: DEBUG nova.network.neutron [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.218074] env[62235]: INFO nova.compute.manager [None req-d6c50e2f-61f0-43c3-9b98-053d6a7f33dd tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] instance snapshotting [ 718.218705] env[62235]: DEBUG nova.objects.instance [None req-d6c50e2f-61f0-43c3-9b98-053d6a7f33dd tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lazy-loading 'flavor' on Instance uuid be4fb088-8b01-4245-82fb-b0e085825597 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 718.289626] env[62235]: DEBUG nova.network.neutron [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.337149] env[62235]: DEBUG nova.network.neutron [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.340016] env[62235]: INFO nova.compute.manager [-] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Took 1.02 seconds to deallocate network for instance. [ 718.342035] env[62235]: DEBUG nova.compute.claims [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 718.342180] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.407895] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.410629] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.473397] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.664274] env[62235]: DEBUG oslo_concurrency.lockutils [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] Releasing lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.664274] env[62235]: DEBUG nova.compute.manager [req-936112fb-f09b-4285-a0ed-291d83ffdd67 req-6ff028a4-ec7b-43ff-bb12-ea6c0982721f service nova] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Received event network-vif-deleted-c0733950-c935-475d-9191-85daffca842b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.719047] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0561ebb-ce60-431e-b002-3e7ea1da3beb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.725554] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3f6016-9ca9-491d-90b2-e9c8e8c20731 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.731645] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577f7158-a9e0-4411-bc86-28433092e915 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.747081] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32744d7-b52e-4719-86b8-412024965830 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.773923] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b83bc7-ee2e-4de0-91d6-c34f1d47ea1d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.784870] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e3ee80-be99-4149-8c4f-1f82640fd419 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.799958] env[62235]: DEBUG nova.compute.provider_tree [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.839983] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Releasing lock "refresh_cache-be4fb088-8b01-4245-82fb-b0e085825597" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.840536] env[62235]: DEBUG nova.compute.manager [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.840751] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.841689] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506c9686-bb76-44f0-9c2f-3aeb7e433cae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.850234] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 718.850574] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd1b9840-0a04-44bd-9e74-b65cc3454234 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.858422] env[62235]: DEBUG oslo_vmware.api [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 718.858422] env[62235]: value = "task-1271478" [ 718.858422] env[62235]: _type = "Task" [ 718.858422] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.866637] env[62235]: DEBUG oslo_vmware.api [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.975751] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Releasing lock "refresh_cache-3f245678-41f4-4685-87fc-822bf877d43e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.976016] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.976197] env[62235]: DEBUG nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.976370] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.000026] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.147461] env[62235]: INFO nova.scheduler.client.report [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Deleted allocations for instance 2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2 [ 719.281627] env[62235]: DEBUG nova.compute.manager [None req-d6c50e2f-61f0-43c3-9b98-053d6a7f33dd tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Instance disappeared during snapshot {{(pid=62235) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 719.301987] env[62235]: DEBUG nova.scheduler.client.report [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.368226] env[62235]: DEBUG oslo_vmware.api [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271478, 'name': PowerOffVM_Task, 'duration_secs': 0.181662} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.368501] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 719.368668] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 719.368907] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8dc85295-98f3-4535-8dfa-b9f0bcad6d30 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.394993] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 719.395258] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 719.395438] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Deleting the datastore file [datastore2] be4fb088-8b01-4245-82fb-b0e085825597 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 719.395690] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c9934e4-0bc4-487a-ac4e-db79165f64cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.402159] env[62235]: DEBUG oslo_vmware.api [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for the task: (returnval){ [ 719.402159] env[62235]: value = "task-1271480" [ 719.402159] env[62235]: _type = "Task" [ 719.402159] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.410739] env[62235]: DEBUG oslo_vmware.api [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.411089] env[62235]: DEBUG nova.compute.manager [None req-d6c50e2f-61f0-43c3-9b98-053d6a7f33dd tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Found 0 images (rotation: 2) {{(pid=62235) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 719.504771] env[62235]: DEBUG nova.network.neutron [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.655753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ff4cef3b-a558-4d87-a207-1c1d46d14264 tempest-ServerMetadataTestJSON-805266811 tempest-ServerMetadataTestJSON-805266811-project-member] Lock "2cc5fd4a-949f-4499-b894-4cfa5c7d2bc2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.018s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.806480] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.807119] env[62235]: ERROR nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Traceback (most recent call last): [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self.driver.spawn(context, instance, image_meta, [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] vm_ref = self.build_virtual_machine(instance, [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.807119] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] for vif in network_info: [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] return self._sync_wrapper(fn, *args, **kwargs) [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self.wait() [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self[:] = self._gt.wait() [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] return self._exit_event.wait() [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] current.throw(*self._exc) [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.807520] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] result = function(*args, **kwargs) [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] return func(*args, **kwargs) [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] raise e [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] nwinfo = self.network_api.allocate_for_instance( [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] created_port_ids = self._update_ports_for_instance( [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] with excutils.save_and_reraise_exception(): [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] self.force_reraise() [ 719.808127] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] raise self.value [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] updated_port = self._update_port( [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] _ensure_no_port_binding_failure(port) [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] raise exception.PortBindingFailed(port_id=port['id']) [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] nova.exception.PortBindingFailed: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. [ 719.808595] env[62235]: ERROR nova.compute.manager [instance: b348bc9f-3bd4-475f-9957-42a99848005c] [ 719.808595] env[62235]: DEBUG nova.compute.utils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.808987] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.260s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.811864] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Build of instance b348bc9f-3bd4-475f-9957-42a99848005c was re-scheduled: Binding failed for port c12de2f6-0783-42e7-a6f1-7c3e08dcae18, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 719.812272] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 719.812511] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.812665] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquired lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.812825] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.913115] env[62235]: DEBUG oslo_vmware.api [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Task: {'id': task-1271480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101814} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.913378] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 719.913574] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 719.913735] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.913982] env[62235]: INFO nova.compute.manager [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Took 1.07 seconds to destroy the instance on the hypervisor. [ 719.914257] env[62235]: DEBUG oslo.service.loopingcall [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.914446] env[62235]: DEBUG nova.compute.manager [-] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.914543] env[62235]: DEBUG nova.network.neutron [-] [instance: be4fb088-8b01-4245-82fb-b0e085825597] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.928947] env[62235]: DEBUG nova.network.neutron [-] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.008414] env[62235]: INFO nova.compute.manager [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] [instance: 3f245678-41f4-4685-87fc-822bf877d43e] Took 1.03 seconds to deallocate network for instance. [ 720.159204] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.353631] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.433683] env[62235]: DEBUG nova.network.neutron [-] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.490250] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.677734] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.710965] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab71c17e-b7f6-4673-b7d0-37551fe3e304 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.719672] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb0a7bf-9ec1-4dee-81ab-f372800c4636 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.752938] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86b7279-573d-4f44-9062-2379f8d1977f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.761106] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e481ee-fca1-4822-9f9b-72e81f8b95fa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.775787] env[62235]: DEBUG nova.compute.provider_tree [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.937555] env[62235]: INFO nova.compute.manager [-] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Took 1.02 seconds to deallocate network for instance. [ 720.992127] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Releasing lock "refresh_cache-b348bc9f-3bd4-475f-9957-42a99848005c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.992127] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 720.992127] env[62235]: DEBUG nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.992246] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.012908] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.060726] env[62235]: INFO nova.scheduler.client.report [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Deleted allocations for instance 3f245678-41f4-4685-87fc-822bf877d43e [ 721.282529] env[62235]: DEBUG nova.scheduler.client.report [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.444332] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.515296] env[62235]: DEBUG nova.network.neutron [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.570102] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8628c309-355c-4f32-893a-7fa2df96605a tempest-MigrationsAdminTest-1999155021 tempest-MigrationsAdminTest-1999155021-project-member] Lock "3f245678-41f4-4685-87fc-822bf877d43e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.678s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.788424] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.788876] env[62235]: ERROR nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Traceback (most recent call last): [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self.driver.spawn(context, instance, image_meta, [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] vm_ref = self.build_virtual_machine(instance, [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.788876] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] for vif in network_info: [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] return self._sync_wrapper(fn, *args, **kwargs) [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self.wait() [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self[:] = self._gt.wait() [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] return self._exit_event.wait() [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] current.throw(*self._exc) [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.790156] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] result = function(*args, **kwargs) [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] return func(*args, **kwargs) [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] raise e [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] nwinfo = self.network_api.allocate_for_instance( [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] created_port_ids = self._update_ports_for_instance( [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] with excutils.save_and_reraise_exception(): [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] self.force_reraise() [ 721.791075] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] raise self.value [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] updated_port = self._update_port( [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] _ensure_no_port_binding_failure(port) [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] raise exception.PortBindingFailed(port_id=port['id']) [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] nova.exception.PortBindingFailed: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. [ 721.791519] env[62235]: ERROR nova.compute.manager [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] [ 721.791519] env[62235]: DEBUG nova.compute.utils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.791800] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Build of instance 19e40767-d9f5-4b78-80f0-7f0f24464d40 was re-scheduled: Binding failed for port 98987fa9-752f-42e8-826b-80ad07983628, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.791800] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.792138] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Acquiring lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.792316] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Acquired lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.792898] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.795270] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.803s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.018622] env[62235]: INFO nova.compute.manager [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: b348bc9f-3bd4-475f-9957-42a99848005c] Took 1.03 seconds to deallocate network for instance. [ 722.072061] env[62235]: DEBUG nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.326530] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.416443] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.594213] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.668148] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41e3326-70b8-48be-bafc-82360e6f22af {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.675698] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508047a2-fed6-485c-840c-5b023b5e64ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.706302] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c5c259-153b-4965-a675-4824668eeae3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.714012] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6533fd8-d148-4d1b-8232-cf871f954e2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.728288] env[62235]: DEBUG nova.compute.provider_tree [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.918221] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Releasing lock "refresh_cache-19e40767-d9f5-4b78-80f0-7f0f24464d40" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.918549] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.918790] env[62235]: DEBUG nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.919041] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.934438] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.051737] env[62235]: INFO nova.scheduler.client.report [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Deleted allocations for instance b348bc9f-3bd4-475f-9957-42a99848005c [ 723.231897] env[62235]: DEBUG nova.scheduler.client.report [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.436936] env[62235]: DEBUG nova.network.neutron [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.563162] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8766445e-a8bc-4885-81a7-f6f672990368 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "b348bc9f-3bd4-475f-9957-42a99848005c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.220s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.738056] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.738445] env[62235]: ERROR nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Traceback (most recent call last): [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self.driver.spawn(context, instance, image_meta, [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] vm_ref = self.build_virtual_machine(instance, [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.738445] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] for vif in network_info: [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] return self._sync_wrapper(fn, *args, **kwargs) [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self.wait() [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self[:] = self._gt.wait() [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] return self._exit_event.wait() [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] current.throw(*self._exc) [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.738819] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] result = function(*args, **kwargs) [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] return func(*args, **kwargs) [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] raise e [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] nwinfo = self.network_api.allocate_for_instance( [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] created_port_ids = self._update_ports_for_instance( [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] with excutils.save_and_reraise_exception(): [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] self.force_reraise() [ 723.739252] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] raise self.value [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] updated_port = self._update_port( [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] _ensure_no_port_binding_failure(port) [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] raise exception.PortBindingFailed(port_id=port['id']) [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] nova.exception.PortBindingFailed: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. [ 723.739656] env[62235]: ERROR nova.compute.manager [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] [ 723.739656] env[62235]: DEBUG nova.compute.utils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.740391] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.307s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.741842] env[62235]: INFO nova.compute.claims [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.744475] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Build of instance 012dd63b-a092-4a1d-a79c-1e1bf2fd8771 was re-scheduled: Binding failed for port ccacb20d-8b17-4df3-afec-4e31be46daa7, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 723.744939] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 723.745181] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquiring lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.745329] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Acquired lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.745485] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.939262] env[62235]: INFO nova.compute.manager [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] [instance: 19e40767-d9f5-4b78-80f0-7f0f24464d40] Took 1.02 seconds to deallocate network for instance. [ 724.065786] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 724.267264] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.358125] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.495238] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "b5a1da09-2137-4434-a1a4-174b196b61b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.495573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.582992] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.861018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Releasing lock "refresh_cache-012dd63b-a092-4a1d-a79c-1e1bf2fd8771" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.861423] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 724.861629] env[62235]: DEBUG nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.861629] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.876891] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.968985] env[62235]: INFO nova.scheduler.client.report [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Deleted allocations for instance 19e40767-d9f5-4b78-80f0-7f0f24464d40 [ 725.081840] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b54cf1-a215-4c14-8e75-543a22f2d340 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.089730] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e7c3e6-0904-4fc3-b76f-a80fe5b84f06 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.120083] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22890ca-8688-4d61-934d-d154f1b2a55b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.127769] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce19cc6-12bb-488b-b49d-86489db7bda6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.141073] env[62235]: DEBUG nova.compute.provider_tree [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.380139] env[62235]: DEBUG nova.network.neutron [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.479539] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0c5a070e-1b56-4c6b-a467-a45e525b25c9 tempest-ServerActionsV293TestJSON-1805265601 tempest-ServerActionsV293TestJSON-1805265601-project-member] Lock "19e40767-d9f5-4b78-80f0-7f0f24464d40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.935s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.644283] env[62235]: DEBUG nova.scheduler.client.report [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.883855] env[62235]: INFO nova.compute.manager [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] [instance: 012dd63b-a092-4a1d-a79c-1e1bf2fd8771] Took 1.02 seconds to deallocate network for instance. [ 725.983986] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.155104] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.155104] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.156290] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.832s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.158063] env[62235]: INFO nova.compute.claims [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.509398] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.667232] env[62235]: DEBUG nova.compute.utils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.670945] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.671170] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.727917] env[62235]: DEBUG nova.policy [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '488eb0da86a2488c8dbf2594fff89272', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f37ce761ef24a59902a405b8accd9d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.920497] env[62235]: INFO nova.scheduler.client.report [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Deleted allocations for instance 012dd63b-a092-4a1d-a79c-1e1bf2fd8771 [ 727.171736] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.416641] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Successfully created port: e9191560-993a-4c95-8d52-d6faede71cf6 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.433118] env[62235]: DEBUG oslo_concurrency.lockutils [None req-667f34a8-2497-4b30-bf7a-36dbf4481427 tempest-ServersAdminTestJSON-435052445 tempest-ServersAdminTestJSON-435052445-project-member] Lock "012dd63b-a092-4a1d-a79c-1e1bf2fd8771" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.433s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.584126] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d720731-7749-443b-b5ff-b110d314c38d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.592013] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d67bbc-7e48-4212-9f1a-da68302b14d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.623765] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6122e4ac-6c89-4fa0-a1a6-8f4c8d307454 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.631085] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2e6512-5845-4003-995c-645a177f5172 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.644605] env[62235]: DEBUG nova.compute.provider_tree [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.939765] env[62235]: DEBUG nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.147373] env[62235]: DEBUG nova.scheduler.client.report [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.183761] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.214264] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.215190] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.215190] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.215190] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.215190] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.216676] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.217602] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.217782] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.218298] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.218298] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.219880] env[62235]: DEBUG nova.virt.hardware [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.221117] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693e4345-431c-4b75-a753-07a38d75f632 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.235016] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e32cb1-e8dc-4c89-9c3e-52108889e6b0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.470558] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.652302] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.652700] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.656075] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.797s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.992230] env[62235]: DEBUG nova.compute.manager [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Received event network-changed-e9191560-993a-4c95-8d52-d6faede71cf6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.992230] env[62235]: DEBUG nova.compute.manager [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Refreshing instance network info cache due to event network-changed-e9191560-993a-4c95-8d52-d6faede71cf6. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.992503] env[62235]: DEBUG oslo_concurrency.lockutils [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] Acquiring lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.992503] env[62235]: DEBUG oslo_concurrency.lockutils [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] Acquired lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.992576] env[62235]: DEBUG nova.network.neutron [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Refreshing network info cache for port e9191560-993a-4c95-8d52-d6faede71cf6 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.164370] env[62235]: DEBUG nova.compute.utils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.165740] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.165912] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 729.269815] env[62235]: ERROR nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. [ 729.269815] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.269815] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.269815] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.269815] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.269815] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.269815] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.269815] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.269815] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.269815] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 729.269815] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.269815] env[62235]: ERROR nova.compute.manager raise self.value [ 729.269815] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.269815] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.269815] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.269815] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.270584] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.270584] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.270584] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. [ 729.270584] env[62235]: ERROR nova.compute.manager [ 729.270584] env[62235]: Traceback (most recent call last): [ 729.270584] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.270584] env[62235]: listener.cb(fileno) [ 729.270584] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.270584] env[62235]: result = function(*args, **kwargs) [ 729.270584] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 729.270584] env[62235]: return func(*args, **kwargs) [ 729.270584] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.270584] env[62235]: raise e [ 729.270584] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.270584] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 729.270584] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.270584] env[62235]: created_port_ids = self._update_ports_for_instance( [ 729.270584] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.270584] env[62235]: with excutils.save_and_reraise_exception(): [ 729.270584] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.270584] env[62235]: self.force_reraise() [ 729.270584] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.270584] env[62235]: raise self.value [ 729.270584] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.270584] env[62235]: updated_port = self._update_port( [ 729.270584] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.270584] env[62235]: _ensure_no_port_binding_failure(port) [ 729.270584] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.270584] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.271756] env[62235]: nova.exception.PortBindingFailed: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. [ 729.271756] env[62235]: Removing descriptor: 16 [ 729.271756] env[62235]: ERROR nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Traceback (most recent call last): [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] yield resources [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self.driver.spawn(context, instance, image_meta, [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.271756] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] vm_ref = self.build_virtual_machine(instance, [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] for vif in network_info: [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return self._sync_wrapper(fn, *args, **kwargs) [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self.wait() [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self[:] = self._gt.wait() [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return self._exit_event.wait() [ 729.272198] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] result = hub.switch() [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return self.greenlet.switch() [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] result = function(*args, **kwargs) [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return func(*args, **kwargs) [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] raise e [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] nwinfo = self.network_api.allocate_for_instance( [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.272695] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] created_port_ids = self._update_ports_for_instance( [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] with excutils.save_and_reraise_exception(): [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self.force_reraise() [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] raise self.value [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] updated_port = self._update_port( [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] _ensure_no_port_binding_failure(port) [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.273273] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] raise exception.PortBindingFailed(port_id=port['id']) [ 729.273852] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] nova.exception.PortBindingFailed: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. [ 729.273852] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] [ 729.273852] env[62235]: INFO nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Terminating instance [ 729.275710] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.289519] env[62235]: DEBUG nova.policy [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'afd81035767e465c97b233d3bcd16fde', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '163d35d07e4b432a977269ab48d5a57c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.517510] env[62235]: DEBUG nova.network.neutron [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.551280] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c35a21-5278-4308-bc3e-368180c226fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.559220] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9050a46b-34cd-4af1-be32-b7cee047e50a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.594338] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7b6f81-8e51-404b-9e04-69116e8f258c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.601859] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955a75e0-5345-43de-98d6-a14687754e01 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.618575] env[62235]: DEBUG nova.compute.provider_tree [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.670954] env[62235]: DEBUG nova.network.neutron [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.672908] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.716008] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Successfully created port: ebe2a32f-c082-4887-bc31-7501887faf6f {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.121938] env[62235]: DEBUG nova.scheduler.client.report [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.180021] env[62235]: DEBUG oslo_concurrency.lockutils [req-09d830d2-1eeb-46b5-bba1-e803b869a55d req-9940b39f-bb06-43c1-887e-865d92e7abec service nova] Releasing lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.180021] env[62235]: INFO nova.virt.block_device [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Booting with volume 8e8acdde-2a50-41b7-96dd-a998fca06f88 at /dev/sda [ 730.180473] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquired lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.180846] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.221011] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-700891a0-435b-4ec1-9477-09ff04488f54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.230250] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78f16f8-79b9-41d4-8559-3ab160e4e559 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.254271] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25cb6e1b-5107-4769-a822-8691525c5087 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.261920] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e515168b-dd82-430e-85c9-366587d7bdd7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.290035] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3973855d-ca67-4a85-b957-df5b38cc9884 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.297527] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b87dfd-4fd6-4179-a074-40790f754bee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.313137] env[62235]: DEBUG nova.virt.block_device [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Updating existing volume attachment record: f6a2d256-6578-4dbc-856e-3a48af1bb8bd {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 730.630968] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.631659] env[62235]: ERROR nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Traceback (most recent call last): [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self.driver.spawn(context, instance, image_meta, [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] vm_ref = self.build_virtual_machine(instance, [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.631659] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] for vif in network_info: [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return self._sync_wrapper(fn, *args, **kwargs) [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self.wait() [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self[:] = self._gt.wait() [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return self._exit_event.wait() [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] result = hub.switch() [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.632151] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return self.greenlet.switch() [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] result = function(*args, **kwargs) [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] return func(*args, **kwargs) [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] raise e [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] nwinfo = self.network_api.allocate_for_instance( [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] created_port_ids = self._update_ports_for_instance( [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] with excutils.save_and_reraise_exception(): [ 730.632652] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] self.force_reraise() [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] raise self.value [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] updated_port = self._update_port( [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] _ensure_no_port_binding_failure(port) [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] raise exception.PortBindingFailed(port_id=port['id']) [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] nova.exception.PortBindingFailed: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. [ 730.633305] env[62235]: ERROR nova.compute.manager [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] [ 730.634833] env[62235]: DEBUG nova.compute.utils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.634833] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.291s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.640020] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Build of instance 693db195-ce52-40ff-bbb4-91f03d5a0bd9 was re-scheduled: Binding failed for port 580a74e8-7dce-4bdd-b04f-aa1f1f68e503, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 730.640020] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 730.640020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Acquiring lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.640020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Acquired lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.640468] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.777060] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.976298] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.029356] env[62235]: DEBUG nova.compute.manager [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Received event network-vif-deleted-e9191560-993a-4c95-8d52-d6faede71cf6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.029617] env[62235]: DEBUG nova.compute.manager [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] [instance: ea4a3f54-388e-451f-9318-78943a169213] Received event network-changed-ebe2a32f-c082-4887-bc31-7501887faf6f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.029808] env[62235]: DEBUG nova.compute.manager [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] [instance: ea4a3f54-388e-451f-9318-78943a169213] Refreshing instance network info cache due to event network-changed-ebe2a32f-c082-4887-bc31-7501887faf6f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 731.030059] env[62235]: DEBUG oslo_concurrency.lockutils [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] Acquiring lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.030228] env[62235]: DEBUG oslo_concurrency.lockutils [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] Acquired lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.030458] env[62235]: DEBUG nova.network.neutron [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] [instance: ea4a3f54-388e-451f-9318-78943a169213] Refreshing network info cache for port ebe2a32f-c082-4887-bc31-7501887faf6f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 731.168648] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.282650] env[62235]: ERROR nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. [ 731.282650] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 731.282650] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.282650] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 731.282650] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.282650] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 731.282650] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.282650] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 731.282650] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.282650] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 731.282650] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.282650] env[62235]: ERROR nova.compute.manager raise self.value [ 731.282650] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.282650] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 731.282650] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.282650] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 731.283619] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.283619] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 731.283619] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. [ 731.283619] env[62235]: ERROR nova.compute.manager [ 731.283619] env[62235]: Traceback (most recent call last): [ 731.283619] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 731.283619] env[62235]: listener.cb(fileno) [ 731.283619] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.283619] env[62235]: result = function(*args, **kwargs) [ 731.283619] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.283619] env[62235]: return func(*args, **kwargs) [ 731.283619] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.283619] env[62235]: raise e [ 731.283619] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.283619] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 731.283619] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.283619] env[62235]: created_port_ids = self._update_ports_for_instance( [ 731.283619] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.283619] env[62235]: with excutils.save_and_reraise_exception(): [ 731.283619] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.283619] env[62235]: self.force_reraise() [ 731.283619] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.283619] env[62235]: raise self.value [ 731.283619] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.283619] env[62235]: updated_port = self._update_port( [ 731.283619] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.283619] env[62235]: _ensure_no_port_binding_failure(port) [ 731.283619] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.283619] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 731.284980] env[62235]: nova.exception.PortBindingFailed: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. [ 731.284980] env[62235]: Removing descriptor: 22 [ 731.380188] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.478886] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Releasing lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.479328] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 731.479543] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.479832] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acf3c650-8142-40d8-ae76-737acd9570bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.489273] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d21732-2462-4e05-a9f6-4e007405f69c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.514554] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c6f4360-b80b-43e3-accc-f7500e626100 could not be found. [ 731.514832] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 731.514957] env[62235]: INFO nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Took 0.04 seconds to destroy the instance on the hypervisor. [ 731.515209] env[62235]: DEBUG oslo.service.loopingcall [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.517928] env[62235]: DEBUG nova.compute.manager [-] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.518047] env[62235]: DEBUG nova.network.neutron [-] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.536978] env[62235]: DEBUG nova.network.neutron [-] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.555462] env[62235]: DEBUG nova.network.neutron [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.563937] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a530a3dc-4dc6-4a20-bd36-dbc1e0f8b0cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.575381] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d01fea-e617-42dd-bed0-6c67aa370434 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.607882] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37722e7-dead-413e-b25f-108e391cde58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.615565] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c8c987-2259-4346-aac7-bb5f197cad48 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.628660] env[62235]: DEBUG nova.compute.provider_tree [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.724268] env[62235]: DEBUG nova.network.neutron [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] [instance: ea4a3f54-388e-451f-9318-78943a169213] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.888219] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Releasing lock "refresh_cache-693db195-ce52-40ff-bbb4-91f03d5a0bd9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.888495] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.888704] env[62235]: DEBUG nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.888816] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.910826] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.040873] env[62235]: DEBUG nova.network.neutron [-] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.132191] env[62235]: DEBUG nova.scheduler.client.report [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.227755] env[62235]: DEBUG oslo_concurrency.lockutils [req-a279d094-5958-4914-adb1-6d3b44d57961 req-2491ad18-9097-4cbc-94d3-ff3726ce1380 service nova] Releasing lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.416627] env[62235]: DEBUG nova.network.neutron [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.441403] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.442084] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.442084] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.442084] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.442279] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.442415] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.442553] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.442746] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.442948] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.443200] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.443412] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.443626] env[62235]: DEBUG nova.virt.hardware [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.444863] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1b1cb6-e9fa-4eb9-b603-5915b964a527 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.454653] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847c6b95-2051-44e9-89a1-c3e9e9c46707 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.473250] env[62235]: ERROR nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] Traceback (most recent call last): [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] yield resources [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self.driver.spawn(context, instance, image_meta, [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] vm_ref = self.build_virtual_machine(instance, [ 732.473250] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] for vif in network_info: [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] return self._sync_wrapper(fn, *args, **kwargs) [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self.wait() [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self[:] = self._gt.wait() [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] return self._exit_event.wait() [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.473579] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] current.throw(*self._exc) [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] result = function(*args, **kwargs) [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] return func(*args, **kwargs) [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] raise e [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] nwinfo = self.network_api.allocate_for_instance( [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] created_port_ids = self._update_ports_for_instance( [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] with excutils.save_and_reraise_exception(): [ 732.474192] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self.force_reraise() [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] raise self.value [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] updated_port = self._update_port( [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] _ensure_no_port_binding_failure(port) [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] raise exception.PortBindingFailed(port_id=port['id']) [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] nova.exception.PortBindingFailed: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. [ 732.474594] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] [ 732.474594] env[62235]: INFO nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Terminating instance [ 732.478952] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Acquiring lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.478952] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Acquired lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.478952] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.543160] env[62235]: INFO nova.compute.manager [-] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Took 1.02 seconds to deallocate network for instance. [ 732.547150] env[62235]: DEBUG nova.compute.claims [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 732.547150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.643022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.643022] env[62235]: ERROR nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. [ 732.643022] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Traceback (most recent call last): [ 732.643022] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.643022] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self.driver.spawn(context, instance, image_meta, [ 732.643022] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 732.643022] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.643022] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.643022] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] vm_ref = self.build_virtual_machine(instance, [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] for vif in network_info: [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return self._sync_wrapper(fn, *args, **kwargs) [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self.wait() [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self[:] = self._gt.wait() [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return self._exit_event.wait() [ 732.643569] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] result = hub.switch() [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return self.greenlet.switch() [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] result = function(*args, **kwargs) [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] return func(*args, **kwargs) [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] raise e [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] nwinfo = self.network_api.allocate_for_instance( [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.644036] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] created_port_ids = self._update_ports_for_instance( [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] with excutils.save_and_reraise_exception(): [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] self.force_reraise() [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] raise self.value [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] updated_port = self._update_port( [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] _ensure_no_port_binding_failure(port) [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.644389] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] raise exception.PortBindingFailed(port_id=port['id']) [ 732.644749] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] nova.exception.PortBindingFailed: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. [ 732.644749] env[62235]: ERROR nova.compute.manager [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] [ 732.644749] env[62235]: DEBUG nova.compute.utils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.644749] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.232s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.644749] env[62235]: INFO nova.compute.claims [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.647062] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Build of instance 452cf04e-d6e9-48a8-a77b-f931c9440858 was re-scheduled: Binding failed for port c0733950-c935-475d-9191-85daffca842b, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.647480] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.647694] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.647844] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquired lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.648124] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.682551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.682551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.921547] env[62235]: INFO nova.compute.manager [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] [instance: 693db195-ce52-40ff-bbb4-91f03d5a0bd9] Took 1.03 seconds to deallocate network for instance. [ 733.000777] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.067703] env[62235]: DEBUG nova.compute.manager [req-afda720c-36c2-40cb-89f4-f9948101dd27 req-326a9c30-74a7-4afe-a68e-eb4723434257 service nova] [instance: ea4a3f54-388e-451f-9318-78943a169213] Received event network-vif-deleted-ebe2a32f-c082-4887-bc31-7501887faf6f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.098136] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.176914] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.272608] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.603140] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Releasing lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.604131] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.604511] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-825f1117-60fc-4142-bf92-aa3208f2ae81 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.613486] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea1ee76-e019-41ef-9137-cb9ebdb5814a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.640023] env[62235]: WARNING nova.virt.vmwareapi.driver [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance ea4a3f54-388e-451f-9318-78943a169213 could not be found. [ 733.640260] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.640529] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27823b51-1dcf-49ea-91b0-d8ec1d55168b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.648088] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6ad865-d252-4864-85ed-6dbd767f2aa3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.671499] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea4a3f54-388e-451f-9318-78943a169213 could not be found. [ 733.671749] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.671935] env[62235]: INFO nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Took 0.07 seconds to destroy the instance on the hypervisor. [ 733.672229] env[62235]: DEBUG oslo.service.loopingcall [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.672513] env[62235]: DEBUG nova.compute.manager [-] [instance: ea4a3f54-388e-451f-9318-78943a169213] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.672627] env[62235]: DEBUG nova.network.neutron [-] [instance: ea4a3f54-388e-451f-9318-78943a169213] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.687902] env[62235]: DEBUG nova.network.neutron [-] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.775482] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Releasing lock "refresh_cache-452cf04e-d6e9-48a8-a77b-f931c9440858" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.775482] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.775609] env[62235]: DEBUG nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.775749] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.793029] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.957244] env[62235]: INFO nova.scheduler.client.report [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Deleted allocations for instance 693db195-ce52-40ff-bbb4-91f03d5a0bd9 [ 733.993975] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7396a82c-5e67-4adb-ae6d-5944e6f0ab26 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.002654] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62737f7e-0be0-4718-9e54-b860a8932f4a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.037883] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86048606-ad98-49d6-8235-a24b05668bab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.047787] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e0e058-c422-4e38-9701-e8b98a83137c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.061891] env[62235]: DEBUG nova.compute.provider_tree [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.190429] env[62235]: DEBUG nova.network.neutron [-] [instance: ea4a3f54-388e-451f-9318-78943a169213] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.297039] env[62235]: DEBUG nova.network.neutron [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.472146] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac73225-6682-4539-8fe2-e76056059968 tempest-ServersTestManualDisk-749583947 tempest-ServersTestManualDisk-749583947-project-member] Lock "693db195-ce52-40ff-bbb4-91f03d5a0bd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.457s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.565363] env[62235]: DEBUG nova.scheduler.client.report [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.693714] env[62235]: INFO nova.compute.manager [-] [instance: ea4a3f54-388e-451f-9318-78943a169213] Took 1.02 seconds to deallocate network for instance. [ 734.800474] env[62235]: INFO nova.compute.manager [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 452cf04e-d6e9-48a8-a77b-f931c9440858] Took 1.02 seconds to deallocate network for instance. [ 734.974882] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 735.071629] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.072192] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.075223] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.398s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.076438] env[62235]: INFO nova.compute.claims [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.254697] env[62235]: INFO nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Took 0.56 seconds to detach 1 volumes for instance. [ 735.256969] env[62235]: DEBUG nova.compute.claims [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.257226] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.502421] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.581435] env[62235]: DEBUG nova.compute.utils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.587604] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.587604] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.626559] env[62235]: DEBUG nova.policy [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '584498528bb64199ab6599a3ff77a116', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d2dd2a7eecd4647a87f9846a8472713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.831830] env[62235]: INFO nova.scheduler.client.report [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Deleted allocations for instance 452cf04e-d6e9-48a8-a77b-f931c9440858 [ 736.092020] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.116129] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Successfully created port: cbee762a-ec5f-4935-b053-b9b3bc7eae0b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.344702] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2489f672-420f-49a4-8607-73de82a1e35e tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "452cf04e-d6e9-48a8-a77b-f931c9440858" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.887s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.472012] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69742bc-843b-460c-a047-5ca5fb5d2fae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.481109] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f580e4-7ea6-48c9-b663-63d9c7d65de7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.516065] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78b509d-59b9-4694-8244-57e86405db7c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.522049] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4104faa3-9142-400d-9da2-d55f368fc13d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.535468] env[62235]: DEBUG nova.compute.provider_tree [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.847424] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 737.031165] env[62235]: DEBUG nova.compute.manager [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Received event network-changed-cbee762a-ec5f-4935-b053-b9b3bc7eae0b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.031165] env[62235]: DEBUG nova.compute.manager [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Refreshing instance network info cache due to event network-changed-cbee762a-ec5f-4935-b053-b9b3bc7eae0b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.031165] env[62235]: DEBUG oslo_concurrency.lockutils [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] Acquiring lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.031165] env[62235]: DEBUG oslo_concurrency.lockutils [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] Acquired lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.031165] env[62235]: DEBUG nova.network.neutron [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Refreshing network info cache for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 737.040148] env[62235]: DEBUG nova.scheduler.client.report [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.106391] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.137670] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.138194] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.138608] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.138939] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.139276] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.139730] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.140519] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.140834] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.143743] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.143743] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.143743] env[62235]: DEBUG nova.virt.hardware [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.143743] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0d1328-6b08-4a7a-bced-ef119f7b5e93 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.154147] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7527079-11ad-4962-a4e1-a47c94ed8e08 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.205551] env[62235]: ERROR nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. [ 737.205551] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 737.205551] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.205551] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 737.205551] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.205551] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 737.205551] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.205551] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 737.205551] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.205551] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 737.205551] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.205551] env[62235]: ERROR nova.compute.manager raise self.value [ 737.205551] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.205551] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 737.205551] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.205551] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 737.206040] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.206040] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 737.206040] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. [ 737.206040] env[62235]: ERROR nova.compute.manager [ 737.206040] env[62235]: Traceback (most recent call last): [ 737.206040] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 737.206040] env[62235]: listener.cb(fileno) [ 737.206040] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.206040] env[62235]: result = function(*args, **kwargs) [ 737.206040] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.206040] env[62235]: return func(*args, **kwargs) [ 737.206040] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.206040] env[62235]: raise e [ 737.206040] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.206040] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 737.206040] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.206040] env[62235]: created_port_ids = self._update_ports_for_instance( [ 737.206040] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.206040] env[62235]: with excutils.save_and_reraise_exception(): [ 737.206040] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.206040] env[62235]: self.force_reraise() [ 737.206040] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.206040] env[62235]: raise self.value [ 737.206040] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.206040] env[62235]: updated_port = self._update_port( [ 737.206040] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.206040] env[62235]: _ensure_no_port_binding_failure(port) [ 737.206040] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.206040] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 737.206932] env[62235]: nova.exception.PortBindingFailed: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. [ 737.206932] env[62235]: Removing descriptor: 22 [ 737.206932] env[62235]: ERROR nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Traceback (most recent call last): [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] yield resources [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self.driver.spawn(context, instance, image_meta, [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.206932] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] vm_ref = self.build_virtual_machine(instance, [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] for vif in network_info: [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return self._sync_wrapper(fn, *args, **kwargs) [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self.wait() [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self[:] = self._gt.wait() [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return self._exit_event.wait() [ 737.207345] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] result = hub.switch() [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return self.greenlet.switch() [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] result = function(*args, **kwargs) [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return func(*args, **kwargs) [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] raise e [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] nwinfo = self.network_api.allocate_for_instance( [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.207738] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] created_port_ids = self._update_ports_for_instance( [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] with excutils.save_and_reraise_exception(): [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self.force_reraise() [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] raise self.value [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] updated_port = self._update_port( [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] _ensure_no_port_binding_failure(port) [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.208214] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] raise exception.PortBindingFailed(port_id=port['id']) [ 737.208614] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] nova.exception.PortBindingFailed: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. [ 737.208614] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] [ 737.208614] env[62235]: INFO nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Terminating instance [ 737.209917] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Acquiring lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.370401] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.543272] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.543812] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.546401] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.102s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.550085] env[62235]: DEBUG nova.objects.instance [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lazy-loading 'resources' on Instance uuid be4fb088-8b01-4245-82fb-b0e085825597 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 737.560363] env[62235]: DEBUG nova.network.neutron [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.687433] env[62235]: DEBUG nova.network.neutron [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.054154] env[62235]: DEBUG nova.compute.utils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.058598] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.058598] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 738.110205] env[62235]: DEBUG nova.policy [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b64f6d6696c042079cced4c83d161f6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4469550d32b4482a9d5e9244cbd681e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.188964] env[62235]: DEBUG oslo_concurrency.lockutils [req-16142888-8ac1-48da-b7d0-224002d9ac0e req-13aa3f7a-4cb1-4682-ab91-ecff9734aa78 service nova] Releasing lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.189872] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Acquired lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.189872] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.396027] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d0e061-afff-43ce-9224-8292ec6080d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.403946] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2507097d-3469-4798-ba0a-019f2c5608c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.440958] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f15e65-c721-480e-b4a8-81a2de385e53 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.448779] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7afc3b-517c-43c7-bbcd-ce9c97a38543 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.468857] env[62235]: DEBUG nova.compute.provider_tree [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.502492] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Successfully created port: 4216d720-b7ac-44ef-8387-21fe8582badc {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.558240] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.722803] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.933106] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.971629] env[62235]: DEBUG nova.scheduler.client.report [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.098290] env[62235]: DEBUG nova.compute.manager [req-cf5df326-834a-441c-b577-c6061fc07eca req-f28c438f-3824-4ea1-8b41-d8360297ada1 service nova] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Received event network-vif-deleted-cbee762a-ec5f-4935-b053-b9b3bc7eae0b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.273940] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.274217] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.433589] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Releasing lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.433966] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.434338] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.436015] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-833c3ab1-3a73-45ee-af74-b1c39655193d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.444197] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a5f857-4a2b-4702-9462-bd8cd892c2fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.465761] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3b4167dd-4535-4482-9452-5548e07dc581 could not be found. [ 739.466013] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.466200] env[62235]: INFO nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Took 0.03 seconds to destroy the instance on the hypervisor. [ 739.466444] env[62235]: DEBUG oslo.service.loopingcall [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.466662] env[62235]: DEBUG nova.compute.manager [-] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.466750] env[62235]: DEBUG nova.network.neutron [-] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.477440] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.478415] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.884s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.479959] env[62235]: INFO nova.compute.claims [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.494389] env[62235]: DEBUG nova.network.neutron [-] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.503748] env[62235]: INFO nova.scheduler.client.report [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Deleted allocations for instance be4fb088-8b01-4245-82fb-b0e085825597 [ 739.568092] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.598410] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.598710] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.598872] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.599067] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.599219] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.599365] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.599573] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.599728] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.599890] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.600061] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.600236] env[62235]: DEBUG nova.virt.hardware [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.601438] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93dab0eb-526a-416b-8e1d-05774fc83b15 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.609510] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a275caa-0f7a-4cf3-9d6d-f4547c3495a6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.657445] env[62235]: ERROR nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. [ 739.657445] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.657445] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.657445] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.657445] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.657445] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.657445] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.657445] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.657445] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.657445] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 739.657445] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.657445] env[62235]: ERROR nova.compute.manager raise self.value [ 739.657445] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.657445] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.657445] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.657445] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.657971] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.657971] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.657971] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. [ 739.657971] env[62235]: ERROR nova.compute.manager [ 739.657971] env[62235]: Traceback (most recent call last): [ 739.657971] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.657971] env[62235]: listener.cb(fileno) [ 739.657971] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.657971] env[62235]: result = function(*args, **kwargs) [ 739.657971] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.657971] env[62235]: return func(*args, **kwargs) [ 739.657971] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.657971] env[62235]: raise e [ 739.657971] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.657971] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 739.657971] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.657971] env[62235]: created_port_ids = self._update_ports_for_instance( [ 739.657971] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.657971] env[62235]: with excutils.save_and_reraise_exception(): [ 739.657971] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.657971] env[62235]: self.force_reraise() [ 739.657971] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.657971] env[62235]: raise self.value [ 739.657971] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.657971] env[62235]: updated_port = self._update_port( [ 739.657971] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.657971] env[62235]: _ensure_no_port_binding_failure(port) [ 739.657971] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.657971] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.658899] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. [ 739.658899] env[62235]: Removing descriptor: 22 [ 739.658899] env[62235]: ERROR nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] Traceback (most recent call last): [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] yield resources [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self.driver.spawn(context, instance, image_meta, [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.658899] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] vm_ref = self.build_virtual_machine(instance, [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] for vif in network_info: [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return self._sync_wrapper(fn, *args, **kwargs) [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self.wait() [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self[:] = self._gt.wait() [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return self._exit_event.wait() [ 739.659311] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] result = hub.switch() [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return self.greenlet.switch() [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] result = function(*args, **kwargs) [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return func(*args, **kwargs) [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] raise e [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] nwinfo = self.network_api.allocate_for_instance( [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.659931] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] created_port_ids = self._update_ports_for_instance( [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] with excutils.save_and_reraise_exception(): [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self.force_reraise() [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] raise self.value [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] updated_port = self._update_port( [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] _ensure_no_port_binding_failure(port) [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.660361] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] raise exception.PortBindingFailed(port_id=port['id']) [ 739.660725] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] nova.exception.PortBindingFailed: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. [ 739.660725] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] [ 739.660725] env[62235]: INFO nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Terminating instance [ 739.661148] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.661398] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.661607] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.996884] env[62235]: DEBUG nova.network.neutron [-] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.012888] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2920e85c-79e0-461e-b3f7-1bdcf3709b26 tempest-ServersAaction247Test-1967172765 tempest-ServersAaction247Test-1967172765-project-member] Lock "be4fb088-8b01-4245-82fb-b0e085825597" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.246s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.183765] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.279225] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.502125] env[62235]: INFO nova.compute.manager [-] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Took 1.03 seconds to deallocate network for instance. [ 740.507376] env[62235]: DEBUG nova.compute.claims [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 740.507376] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.782615] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.784492] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.784685] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.785184] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0fb87c02-8fa0-4f7b-aa62-1211577ef911 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.796422] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f25913-848c-452a-b23a-84717efe3965 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.820531] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 86e61430-ce9a-436e-9331-6276604610e3 could not be found. [ 740.820768] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.820950] env[62235]: INFO nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 740.821300] env[62235]: DEBUG oslo.service.loopingcall [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.825338] env[62235]: DEBUG nova.compute.manager [-] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.825489] env[62235]: DEBUG nova.network.neutron [-] [instance: 86e61430-ce9a-436e-9331-6276604610e3] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.859676] env[62235]: DEBUG nova.network.neutron [-] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.910215] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73970817-1038-4446-87f5-f88834ac68e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.922243] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046d9c45-bc6c-498a-afd9-8ba5c8b8c0e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.956314] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d137ae55-37df-472f-8055-f030734d8496 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.963858] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0759d6bf-2edb-4674-8b5b-f605d0b0b015 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.979151] env[62235]: DEBUG nova.compute.provider_tree [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.179810] env[62235]: DEBUG nova.compute.manager [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Received event network-changed-4216d720-b7ac-44ef-8387-21fe8582badc {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.179987] env[62235]: DEBUG nova.compute.manager [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Refreshing instance network info cache due to event network-changed-4216d720-b7ac-44ef-8387-21fe8582badc. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 741.180244] env[62235]: DEBUG oslo_concurrency.lockutils [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] Acquiring lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.180432] env[62235]: DEBUG oslo_concurrency.lockutils [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] Acquired lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.181929] env[62235]: DEBUG nova.network.neutron [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Refreshing network info cache for port 4216d720-b7ac-44ef-8387-21fe8582badc {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 741.366988] env[62235]: DEBUG nova.network.neutron [-] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.481600] env[62235]: DEBUG nova.scheduler.client.report [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.704671] env[62235]: DEBUG nova.network.neutron [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.849152] env[62235]: DEBUG nova.network.neutron [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.869784] env[62235]: INFO nova.compute.manager [-] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Took 1.04 seconds to deallocate network for instance. [ 741.874579] env[62235]: DEBUG nova.compute.claims [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.874579] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.986340] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.986871] env[62235]: DEBUG nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.990028] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.407s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.991884] env[62235]: INFO nova.compute.claims [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.352263] env[62235]: DEBUG oslo_concurrency.lockutils [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] Releasing lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.352523] env[62235]: DEBUG nova.compute.manager [req-1cf3f93b-2a8f-4f8e-9abf-ab94e38749e8 req-4a50caac-e4da-4b66-8873-3bf424051d75 service nova] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Received event network-vif-deleted-4216d720-b7ac-44ef-8387-21fe8582badc {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.493469] env[62235]: DEBUG nova.compute.utils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.493469] env[62235]: DEBUG nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Not allocating networking since 'none' was specified. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 742.995429] env[62235]: DEBUG nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.326507] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926db3fe-8b6f-4377-8aa6-29ab85e9bfd6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.334327] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904ee778-2a9f-4b95-9c1b-bc3e1ba5fd75 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.365783] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd1e845-8db7-48d3-ae2b-c96064c39d1b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.375027] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d61066-2c45-4748-b685-c314892daf0a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.385933] env[62235]: DEBUG nova.compute.provider_tree [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.890090] env[62235]: DEBUG nova.scheduler.client.report [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.006351] env[62235]: DEBUG nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 744.030475] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.030724] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.030880] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.031088] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.031260] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.031410] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.031624] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.031762] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.031918] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.032107] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.032280] env[62235]: DEBUG nova.virt.hardware [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.033151] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760a7756-79ea-438d-8c6a-137b312630c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.041324] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c3c0b5-973f-4fa0-8d96-e80e4de20e00 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.054446] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.060075] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Creating folder: Project (6823d2b4446841c8881469c9e77d3df7). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 744.060309] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35deb958-896d-4430-a8c3-dee2776f1e31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.071194] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Created folder: Project (6823d2b4446841c8881469c9e77d3df7) in parent group-v273362. [ 744.071376] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Creating folder: Instances. Parent ref: group-v273382. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 744.071590] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c66cd43b-c59a-4829-875b-6dacb7014de9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.080393] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Created folder: Instances in parent group-v273382. [ 744.080621] env[62235]: DEBUG oslo.service.loopingcall [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.080799] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 744.080980] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0be2e584-699d-4ea8-8cfa-a153664bb2f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.096844] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.096844] env[62235]: value = "task-1271484" [ 744.096844] env[62235]: _type = "Task" [ 744.096844] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.103812] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271484, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.395710] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.396265] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.398899] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.890s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.400952] env[62235]: INFO nova.compute.claims [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.606924] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271484, 'name': CreateVM_Task, 'duration_secs': 0.240629} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.607164] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 744.607636] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.607841] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.608174] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 744.608418] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-919810cc-84b2-4d16-a922-db3928cdbdc7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.612500] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 744.612500] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526a1a40-ceb9-a97d-5a75-0f54dd39148b" [ 744.612500] env[62235]: _type = "Task" [ 744.612500] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.620067] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526a1a40-ceb9-a97d-5a75-0f54dd39148b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.905478] env[62235]: DEBUG nova.compute.utils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.908705] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.908872] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 744.960827] env[62235]: DEBUG nova.policy [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f6f9cd7af6784e468ff9e3230fdb433f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d1081d5b610424d833a5f7611c1e796', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.123449] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526a1a40-ceb9-a97d-5a75-0f54dd39148b, 'name': SearchDatastore_Task, 'duration_secs': 0.009253} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.126880] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.126880] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.126880] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.126880] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.126880] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.127226] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83adc029-490f-42aa-a2f0-496d41322f24 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.140410] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.140595] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 745.141349] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da68892c-79df-45b0-b2c3-ba7c47fda5be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.147777] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 745.147777] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a53076-4f7d-2903-9ec0-a1d3f1a21d50" [ 745.147777] env[62235]: _type = "Task" [ 745.147777] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.157588] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a53076-4f7d-2903-9ec0-a1d3f1a21d50, 'name': SearchDatastore_Task, 'duration_secs': 0.007211} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.161671] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4c5352a-bef6-4353-a897-ed3646d324bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.166507] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 745.166507] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523d412a-d715-64ec-c46d-6aa5e113f19f" [ 745.166507] env[62235]: _type = "Task" [ 745.166507] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.173817] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523d412a-d715-64ec-c46d-6aa5e113f19f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.254624] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Successfully created port: 524e4b31-e355-46e2-b753-4d5e24d4151b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.413020] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.677305] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523d412a-d715-64ec-c46d-6aa5e113f19f, 'name': SearchDatastore_Task, 'duration_secs': 0.007852} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.677569] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.678384] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 745.678384] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b94d404e-bd1c-4fad-a7f8-7eec8cd3cf0b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.684428] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 745.684428] env[62235]: value = "task-1271485" [ 745.684428] env[62235]: _type = "Task" [ 745.684428] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.693988] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271485, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.742506] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2016236b-0824-4178-a603-cee2541226fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.749752] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd832b26-c2d6-47a9-b42f-4b8fbe739b4b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.782088] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f08b67-08aa-4336-9a19-340f89a62816 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.789803] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5bca72-c3d5-4343-9727-b8c4abea4fdf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.803859] env[62235]: DEBUG nova.compute.provider_tree [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.195391] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271485, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442643} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.196029] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 746.199023] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.199023] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f47ba64c-2695-461a-8091-2472ef972ae4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.205810] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 746.205810] env[62235]: value = "task-1271486" [ 746.205810] env[62235]: _type = "Task" [ 746.205810] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.214284] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271486, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.226178] env[62235]: DEBUG nova.compute.manager [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Received event network-changed-524e4b31-e355-46e2-b753-4d5e24d4151b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.226178] env[62235]: DEBUG nova.compute.manager [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Refreshing instance network info cache due to event network-changed-524e4b31-e355-46e2-b753-4d5e24d4151b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.226178] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] Acquiring lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.226178] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] Acquired lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.226178] env[62235]: DEBUG nova.network.neutron [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Refreshing network info cache for port 524e4b31-e355-46e2-b753-4d5e24d4151b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 746.310142] env[62235]: DEBUG nova.scheduler.client.report [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.407348] env[62235]: ERROR nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. [ 746.407348] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.407348] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.407348] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.407348] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.407348] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.407348] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.407348] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.407348] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.407348] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 746.407348] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.407348] env[62235]: ERROR nova.compute.manager raise self.value [ 746.407348] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.407348] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.407348] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.407348] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.407924] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.407924] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.407924] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. [ 746.407924] env[62235]: ERROR nova.compute.manager [ 746.407924] env[62235]: Traceback (most recent call last): [ 746.407924] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.407924] env[62235]: listener.cb(fileno) [ 746.407924] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.407924] env[62235]: result = function(*args, **kwargs) [ 746.407924] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.407924] env[62235]: return func(*args, **kwargs) [ 746.407924] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.407924] env[62235]: raise e [ 746.407924] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.407924] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 746.407924] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.407924] env[62235]: created_port_ids = self._update_ports_for_instance( [ 746.407924] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.407924] env[62235]: with excutils.save_and_reraise_exception(): [ 746.407924] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.407924] env[62235]: self.force_reraise() [ 746.407924] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.407924] env[62235]: raise self.value [ 746.407924] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.407924] env[62235]: updated_port = self._update_port( [ 746.407924] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.407924] env[62235]: _ensure_no_port_binding_failure(port) [ 746.407924] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.407924] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.408881] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. [ 746.408881] env[62235]: Removing descriptor: 22 [ 746.421860] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.446316] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.446566] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.446722] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.446904] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.447061] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.447229] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.447450] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.447607] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.447770] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.447927] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.448107] env[62235]: DEBUG nova.virt.hardware [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.448947] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03728ef2-38a9-4f06-9eef-e53435ac0d42 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.457490] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6268cd-a7b2-47c6-93e8-4da49540ae65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.470810] env[62235]: ERROR nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Traceback (most recent call last): [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] yield resources [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self.driver.spawn(context, instance, image_meta, [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] vm_ref = self.build_virtual_machine(instance, [ 746.470810] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] for vif in network_info: [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] return self._sync_wrapper(fn, *args, **kwargs) [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self.wait() [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self[:] = self._gt.wait() [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] return self._exit_event.wait() [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.471209] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] current.throw(*self._exc) [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] result = function(*args, **kwargs) [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] return func(*args, **kwargs) [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] raise e [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] nwinfo = self.network_api.allocate_for_instance( [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] created_port_ids = self._update_ports_for_instance( [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] with excutils.save_and_reraise_exception(): [ 746.471661] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self.force_reraise() [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] raise self.value [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] updated_port = self._update_port( [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] _ensure_no_port_binding_failure(port) [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] raise exception.PortBindingFailed(port_id=port['id']) [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] nova.exception.PortBindingFailed: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. [ 746.472126] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] [ 746.472126] env[62235]: INFO nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Terminating instance [ 746.473696] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.717764] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271486, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062644} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.718076] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.719008] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0989361-c7d1-4c9f-a682-c81a09c0e72c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.740913] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.741190] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecb45af0-7a68-440a-8b2b-2716928f6ba8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.763037] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 746.763037] env[62235]: value = "task-1271487" [ 746.763037] env[62235]: _type = "Task" [ 746.763037] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.766984] env[62235]: DEBUG nova.network.neutron [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.771746] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271487, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.815210] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.815802] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.818324] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.348s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.819742] env[62235]: INFO nova.compute.claims [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.852124] env[62235]: DEBUG nova.network.neutron [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.272476] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271487, 'name': ReconfigVM_Task, 'duration_secs': 0.383697} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.272778] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.273384] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4cf3d048-9b7d-4fd1-b8f7-8cedd439c31b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.279504] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 747.279504] env[62235]: value = "task-1271488" [ 747.279504] env[62235]: _type = "Task" [ 747.279504] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.286739] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271488, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.324433] env[62235]: DEBUG nova.compute.utils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.328075] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.328265] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.354964] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b3cbf8e-1063-407a-8beb-9e180286f6ee req-ba3499d8-9f36-447e-a50a-332006f9f8c8 service nova] Releasing lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.355584] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquired lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.355584] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.383015] env[62235]: DEBUG nova.policy [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5ec091a9504485bba6048fce1a1d090', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91d149e97a7b4bdfa43ed32cb59d0fdf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.638732] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Successfully created port: 932dd0ad-9494-4ebf-b0c0-ff87b847e67d {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.793948] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271488, 'name': Rename_Task, 'duration_secs': 0.136213} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.794287] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 747.794548] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-652e971f-b93b-4e82-b3c8-8e7dea075432 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.801554] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 747.801554] env[62235]: value = "task-1271489" [ 747.801554] env[62235]: _type = "Task" [ 747.801554] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.810628] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.829339] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.892836] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.957349] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Successfully created port: 09fdfd57-69bd-44cf-917f-facc7cccc287 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.029031] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.160970] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34ef735-dd05-4ccb-9828-5d1db1b26c79 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.168734] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 748.169043] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 748.171177] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e5c537-42b2-4801-9971-202e5c8ea89c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.207385] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0001772f-7ac0-4594-9ad9-5980bffd2d2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.215568] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea5ca3c-0ba8-431d-b3c1-386c2bab47bb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.229283] env[62235]: DEBUG nova.compute.provider_tree [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.270192] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Successfully created port: bc9a1265-500f-407d-8c1c-10583009e1e5 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.289481] env[62235]: DEBUG nova.compute.manager [req-7b1ade50-e781-4dcc-b02f-c6a1d43663de req-1f0eb1bf-f5ce-4453-9d69-444c0826e198 service nova] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Received event network-vif-deleted-524e4b31-e355-46e2-b753-4d5e24d4151b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.312252] env[62235]: DEBUG oslo_vmware.api [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271489, 'name': PowerOnVM_Task, 'duration_secs': 0.400207} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.312515] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 748.312706] env[62235]: INFO nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Took 4.31 seconds to spawn the instance on the hypervisor. [ 748.312879] env[62235]: DEBUG nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 748.313652] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbaa00d-cff6-4dbc-af9d-9744c5c187b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.532834] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Releasing lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.532834] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.532834] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.533591] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23ac6d1e-daa4-4eb0-ad3c-372abb66ab1a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.542889] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b177919e-b437-4051-91a6-3ba435fa159c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.565851] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c62172db-7e1c-4fd1-b8b9-a8daf0a71f91 could not be found. [ 748.566091] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.566274] env[62235]: INFO nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Took 0.03 seconds to destroy the instance on the hypervisor. [ 748.566517] env[62235]: DEBUG oslo.service.loopingcall [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.566729] env[62235]: DEBUG nova.compute.manager [-] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.566820] env[62235]: DEBUG nova.network.neutron [-] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.588010] env[62235]: DEBUG nova.network.neutron [-] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.681627] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 748.681800] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 748.681920] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 748.732992] env[62235]: DEBUG nova.scheduler.client.report [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.831197] env[62235]: INFO nova.compute.manager [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Took 26.25 seconds to build instance. [ 748.843031] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.874361] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.874612] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.874789] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.875069] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.875248] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.875414] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.875656] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.875882] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.876020] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.876242] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.876482] env[62235]: DEBUG nova.virt.hardware [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.877389] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a738112-08e8-4d87-bb06-6daa41f99eaa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.887308] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b313a4-0dcd-47be-be3c-7d2753ecd064 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.088730] env[62235]: DEBUG nova.network.neutron [-] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.190571] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 749.190571] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ea4a3f54-388e-451f-9318-78943a169213] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 749.190571] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 749.190571] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 749.190571] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 749.190571] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 749.190776] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 749.208912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-430056ca-ccba-45c5-adf7-9c068f81f9ec" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.208912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-430056ca-ccba-45c5-adf7-9c068f81f9ec" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.208912] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 749.208912] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid 430056ca-ccba-45c5-adf7-9c068f81f9ec {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 749.240117] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.240541] env[62235]: DEBUG nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 749.243244] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.696s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.301631] env[62235]: ERROR nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. [ 749.301631] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.301631] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.301631] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.301631] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.301631] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.301631] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.301631] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.301631] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.301631] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 749.301631] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.301631] env[62235]: ERROR nova.compute.manager raise self.value [ 749.301631] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.301631] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.301631] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.301631] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.302444] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.302444] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.302444] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. [ 749.302444] env[62235]: ERROR nova.compute.manager [ 749.302444] env[62235]: Traceback (most recent call last): [ 749.302444] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.302444] env[62235]: listener.cb(fileno) [ 749.302444] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.302444] env[62235]: result = function(*args, **kwargs) [ 749.302444] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.302444] env[62235]: return func(*args, **kwargs) [ 749.302444] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.302444] env[62235]: raise e [ 749.302444] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.302444] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 749.302444] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.302444] env[62235]: created_port_ids = self._update_ports_for_instance( [ 749.302444] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.302444] env[62235]: with excutils.save_and_reraise_exception(): [ 749.302444] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.302444] env[62235]: self.force_reraise() [ 749.302444] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.302444] env[62235]: raise self.value [ 749.302444] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.302444] env[62235]: updated_port = self._update_port( [ 749.302444] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.302444] env[62235]: _ensure_no_port_binding_failure(port) [ 749.302444] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.302444] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.303529] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. [ 749.303529] env[62235]: Removing descriptor: 22 [ 749.303529] env[62235]: ERROR nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Traceback (most recent call last): [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] yield resources [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self.driver.spawn(context, instance, image_meta, [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.303529] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] vm_ref = self.build_virtual_machine(instance, [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] for vif in network_info: [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return self._sync_wrapper(fn, *args, **kwargs) [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self.wait() [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self[:] = self._gt.wait() [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return self._exit_event.wait() [ 749.304497] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] result = hub.switch() [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return self.greenlet.switch() [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] result = function(*args, **kwargs) [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return func(*args, **kwargs) [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] raise e [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] nwinfo = self.network_api.allocate_for_instance( [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.305118] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] created_port_ids = self._update_ports_for_instance( [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] with excutils.save_and_reraise_exception(): [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self.force_reraise() [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] raise self.value [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] updated_port = self._update_port( [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] _ensure_no_port_binding_failure(port) [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.305719] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] raise exception.PortBindingFailed(port_id=port['id']) [ 749.306152] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] nova.exception.PortBindingFailed: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. [ 749.306152] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] [ 749.306152] env[62235]: INFO nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Terminating instance [ 749.306152] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.306152] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquired lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.306152] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.333763] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4b403ee9-5716-48be-9adf-b538c9e680aa tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "430056ca-ccba-45c5-adf7-9c068f81f9ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.078s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.492079] env[62235]: INFO nova.compute.manager [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Rebuilding instance [ 749.536309] env[62235]: DEBUG nova.compute.manager [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.537240] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce366cd-8875-40cd-a2ad-1da3619489ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.593256] env[62235]: INFO nova.compute.manager [-] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Took 1.03 seconds to deallocate network for instance. [ 749.595343] env[62235]: DEBUG nova.compute.claims [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 749.595516] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.747532] env[62235]: DEBUG nova.compute.utils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.751647] env[62235]: DEBUG nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Not allocating networking since 'none' was specified. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 749.836084] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.859293] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.948014] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.037644] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcc3669-a453-47f6-b7b9-b1b07ce40f29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.045153] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240b8181-ac7d-4b68-9dd2-fc3d9baf4381 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.050581] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 750.050581] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28aa11af-9469-460c-96eb-9786bef65293 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.091769] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e705ffbe-eb98-44b6-9da7-68549e3c9c87 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.094805] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 750.094805] env[62235]: value = "task-1271490" [ 750.094805] env[62235]: _type = "Task" [ 750.094805] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.101865] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02f4b50-9362-4100-9d72-b770e57a1e0c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.109298] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271490, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.119632] env[62235]: DEBUG nova.compute.provider_tree [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.235381] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.252138] env[62235]: DEBUG nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 750.315748] env[62235]: DEBUG nova.compute.manager [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Received event network-changed-932dd0ad-9494-4ebf-b0c0-ff87b847e67d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.315993] env[62235]: DEBUG nova.compute.manager [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Refreshing instance network info cache due to event network-changed-932dd0ad-9494-4ebf-b0c0-ff87b847e67d. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.316165] env[62235]: DEBUG oslo_concurrency.lockutils [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] Acquiring lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.358895] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.450738] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Releasing lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.450738] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 750.450738] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 750.451043] env[62235]: DEBUG oslo_concurrency.lockutils [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] Acquired lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.451214] env[62235]: DEBUG nova.network.neutron [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Refreshing network info cache for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.453936] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cabe15a9-e8d1-438f-9f71-4ef554fc8a77 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.467176] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfadc53-748b-43b5-a0cd-7d8d3d2a3f4a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.492733] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7e6ebba9-58e8-4971-b9fb-8365a61d5505 could not be found. [ 750.492733] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 750.492733] env[62235]: INFO nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Took 0.04 seconds to destroy the instance on the hypervisor. [ 750.492733] env[62235]: DEBUG oslo.service.loopingcall [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.492733] env[62235]: DEBUG nova.compute.manager [-] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.493049] env[62235]: DEBUG nova.network.neutron [-] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.552497] env[62235]: DEBUG nova.network.neutron [-] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.604728] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271490, 'name': PowerOffVM_Task, 'duration_secs': 0.174323} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.604912] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 750.606033] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 750.606150] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aaa6b2d-3895-4a77-8d03-3d9b5160dcbf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.612899] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 750.613137] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-082a5d52-4603-4140-9058-955fdb6f8f33 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.628018] env[62235]: DEBUG nova.scheduler.client.report [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.637237] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 750.637472] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 750.637717] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Deleting the datastore file [datastore1] 430056ca-ccba-45c5-adf7-9c068f81f9ec {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 750.637898] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-381e5886-da6b-454c-9936-560d6f7142d7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.652600] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 750.652600] env[62235]: value = "task-1271492" [ 750.652600] env[62235]: _type = "Task" [ 750.652600] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.661212] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271492, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.825210] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.977438] env[62235]: DEBUG nova.network.neutron [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.078037] env[62235]: DEBUG nova.network.neutron [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.134349] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.891s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.135027] env[62235]: ERROR nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Traceback (most recent call last): [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self.driver.spawn(context, instance, image_meta, [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] vm_ref = self.build_virtual_machine(instance, [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.135027] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] for vif in network_info: [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return self._sync_wrapper(fn, *args, **kwargs) [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self.wait() [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self[:] = self._gt.wait() [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return self._exit_event.wait() [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] result = hub.switch() [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.135379] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return self.greenlet.switch() [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] result = function(*args, **kwargs) [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] return func(*args, **kwargs) [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] raise e [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] nwinfo = self.network_api.allocate_for_instance( [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] created_port_ids = self._update_ports_for_instance( [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] with excutils.save_and_reraise_exception(): [ 751.135799] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] self.force_reraise() [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] raise self.value [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] updated_port = self._update_port( [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] _ensure_no_port_binding_failure(port) [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] raise exception.PortBindingFailed(port_id=port['id']) [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] nova.exception.PortBindingFailed: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. [ 751.136177] env[62235]: ERROR nova.compute.manager [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] [ 751.136482] env[62235]: DEBUG nova.compute.utils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.136951] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.880s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.139928] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Build of instance 1c6f4360-b80b-43e3-accc-f7500e626100 was re-scheduled: Binding failed for port e9191560-993a-4c95-8d52-d6faede71cf6, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 751.140371] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 751.140594] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquiring lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.140739] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Acquired lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.140897] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.162985] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271492, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095706} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.163323] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 751.163547] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 751.163727] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.263130] env[62235]: DEBUG nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 751.287515] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 751.287515] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 751.287515] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.287758] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 751.287819] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.287967] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 751.288199] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 751.288358] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 751.288524] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 751.288684] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 751.288850] env[62235]: DEBUG nova.virt.hardware [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.289733] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a06da73-c7e3-49d9-aee6-639071e79820 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.297801] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf6c626-704b-45c3-af84-dc5e3cfa6e4d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.310840] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.316332] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Creating folder: Project (0e964fdfba3c4cef999d894f776fee56). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.316622] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8693f326-252d-4c1b-858a-ea8130513ff6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.326945] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-430056ca-ccba-45c5-adf7-9c068f81f9ec" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.327147] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 751.327365] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Created folder: Project (0e964fdfba3c4cef999d894f776fee56) in parent group-v273362. [ 751.327556] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Creating folder: Instances. Parent ref: group-v273385. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.327745] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 751.327934] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b384e38-970e-48f8-b554-25f99d1b399a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.329433] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 751.330019] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 751.330416] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 751.330581] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 751.330732] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 751.330859] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 751.330999] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 751.338470] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Created folder: Instances in parent group-v273385. [ 751.338470] env[62235]: DEBUG oslo.service.loopingcall [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.338470] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 751.338470] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e09b4c6-f0a8-43cc-bc6b-47f286697c1f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.355798] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.355798] env[62235]: value = "task-1271495" [ 751.355798] env[62235]: _type = "Task" [ 751.355798] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.363337] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271495, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.519810] env[62235]: DEBUG nova.network.neutron [-] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.580882] env[62235]: DEBUG oslo_concurrency.lockutils [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] Releasing lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.581196] env[62235]: DEBUG nova.compute.manager [req-6dc75532-5837-409b-9a3a-0ec6e4d1eced req-6918aa9e-5ca9-417e-af05-f3ac52fea5b2 service nova] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Received event network-vif-deleted-932dd0ad-9494-4ebf-b0c0-ff87b847e67d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.660754] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.764909] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.836389] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.867953] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271495, 'name': CreateVM_Task, 'duration_secs': 0.23379} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.868146] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 751.868641] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.868815] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.869351] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 751.869609] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-169bafdd-7d99-44dc-a580-b97f2423dce3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.877536] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 751.877536] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cd3cbf-bad7-8540-2237-eeb0aa8abde1" [ 751.877536] env[62235]: _type = "Task" [ 751.877536] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.885723] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cd3cbf-bad7-8540-2237-eeb0aa8abde1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.940602] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56370a7f-2f22-44ac-9018-99aace3068ff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.947713] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ca66d6-99cd-4502-9d4c-d7e08412294a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.978101] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347ec065-8071-445e-ad13-6dff5b3a6ab6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.985274] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8ca50b-1d8a-4620-8a85-a32cb9047e54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.998964] env[62235]: DEBUG nova.compute.provider_tree [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.022473] env[62235]: INFO nova.compute.manager [-] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Took 1.53 seconds to deallocate network for instance. [ 752.024661] env[62235]: DEBUG nova.compute.claims [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.024846] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.193090] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.193347] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.193580] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.193822] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.193973] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.194138] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.194347] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.194546] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.194733] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.194898] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.195083] env[62235]: DEBUG nova.virt.hardware [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.195922] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fcb0d0-4a93-41cb-99b8-538f53907e42 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.204211] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fe708d-3755-4c76-8be2-d1cc510a773b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.217893] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.223455] env[62235]: DEBUG oslo.service.loopingcall [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.223659] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.223862] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38cd210e-7226-415d-924a-61cd5389e0e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.240213] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.240213] env[62235]: value = "task-1271496" [ 752.240213] env[62235]: _type = "Task" [ 752.240213] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.247421] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271496, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.267433] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Releasing lock "refresh_cache-1c6f4360-b80b-43e3-accc-f7500e626100" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.267666] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 752.267857] env[62235]: DEBUG nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.268040] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.284060] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.389967] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cd3cbf-bad7-8540-2237-eeb0aa8abde1, 'name': SearchDatastore_Task, 'duration_secs': 0.008829} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.389967] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.390365] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.390518] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.390655] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.390968] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.391348] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b55f5798-2b5d-4e03-ad95-0b0de09bcbc1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.402238] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.402238] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 752.402238] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cdf1703-0e3a-4dff-b0ed-57c785a3ae61 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.408773] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 752.408773] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525a2bfe-a08e-050a-b823-eb3e3abf85e9" [ 752.408773] env[62235]: _type = "Task" [ 752.408773] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.416679] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525a2bfe-a08e-050a-b823-eb3e3abf85e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.501700] env[62235]: DEBUG nova.scheduler.client.report [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.750367] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271496, 'name': CreateVM_Task, 'duration_secs': 0.257985} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.750538] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.750943] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.751111] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.751409] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.751638] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c3a9d15-72e3-4eac-b173-90fc4a5291d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.755845] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 752.755845] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5209ca7a-6f5b-b3d6-bbd8-53f198761546" [ 752.755845] env[62235]: _type = "Task" [ 752.755845] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.762976] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5209ca7a-6f5b-b3d6-bbd8-53f198761546, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.787233] env[62235]: DEBUG nova.network.neutron [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.919040] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525a2bfe-a08e-050a-b823-eb3e3abf85e9, 'name': SearchDatastore_Task, 'duration_secs': 0.008446} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.919806] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e71ad424-b110-4af2-88d5-e1f480623a27 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.924569] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 752.924569] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523b7968-7c80-09db-6b2a-fc965b97e62b" [ 752.924569] env[62235]: _type = "Task" [ 752.924569] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.931859] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523b7968-7c80-09db-6b2a-fc965b97e62b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.007099] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.007736] env[62235]: ERROR nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] Traceback (most recent call last): [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self.driver.spawn(context, instance, image_meta, [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] vm_ref = self.build_virtual_machine(instance, [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.007736] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] for vif in network_info: [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] return self._sync_wrapper(fn, *args, **kwargs) [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self.wait() [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self[:] = self._gt.wait() [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] return self._exit_event.wait() [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] current.throw(*self._exc) [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.008132] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] result = function(*args, **kwargs) [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] return func(*args, **kwargs) [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] raise e [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] nwinfo = self.network_api.allocate_for_instance( [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] created_port_ids = self._update_ports_for_instance( [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] with excutils.save_and_reraise_exception(): [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] self.force_reraise() [ 753.008602] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] raise self.value [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] updated_port = self._update_port( [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] _ensure_no_port_binding_failure(port) [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] raise exception.PortBindingFailed(port_id=port['id']) [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] nova.exception.PortBindingFailed: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. [ 753.009261] env[62235]: ERROR nova.compute.manager [instance: ea4a3f54-388e-451f-9318-78943a169213] [ 753.009261] env[62235]: DEBUG nova.compute.utils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.009832] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.508s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.011602] env[62235]: INFO nova.compute.claims [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.014569] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Build of instance ea4a3f54-388e-451f-9318-78943a169213 was re-scheduled: Binding failed for port ebe2a32f-c082-4887-bc31-7501887faf6f, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.015018] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 753.015251] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Acquiring lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.015457] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Acquired lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.015558] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.266225] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5209ca7a-6f5b-b3d6-bbd8-53f198761546, 'name': SearchDatastore_Task, 'duration_secs': 0.009045} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.266543] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.266777] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.266984] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.289665] env[62235]: INFO nova.compute.manager [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] [instance: 1c6f4360-b80b-43e3-accc-f7500e626100] Took 1.02 seconds to deallocate network for instance. [ 753.435032] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523b7968-7c80-09db-6b2a-fc965b97e62b, 'name': SearchDatastore_Task, 'duration_secs': 0.008823} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.435308] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.435539] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 753.435836] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.436030] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.436248] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06a3382e-d6d4-4ada-81cd-ce7614121918 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.438257] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-647fe888-b104-48dc-9875-987f06532065 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.445372] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 753.445372] env[62235]: value = "task-1271497" [ 753.445372] env[62235]: _type = "Task" [ 753.445372] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.446487] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.446667] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 753.450084] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f55351ae-5479-48a7-a125-914f46b2fdeb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.455932] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 753.455932] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f3da68-1b62-feb7-fb81-d2402a73bccb" [ 753.455932] env[62235]: _type = "Task" [ 753.455932] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.458642] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271497, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.465853] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f3da68-1b62-feb7-fb81-d2402a73bccb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.535383] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.608962] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.958048] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271497, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434947} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.958048] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 753.958048] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.958293] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-344732eb-64e7-4a87-aee8-15d5a12c6d2e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.968420] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f3da68-1b62-feb7-fb81-d2402a73bccb, 'name': SearchDatastore_Task, 'duration_secs': 0.015577} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.970056] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 753.970056] env[62235]: value = "task-1271498" [ 753.970056] env[62235]: _type = "Task" [ 753.970056] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.970316] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f37d020-f6d6-4f38-8dc6-740401f5c385 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.977892] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 753.977892] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521c1925-127f-0651-e167-d9ce1b4c928f" [ 753.977892] env[62235]: _type = "Task" [ 753.977892] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.980755] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271498, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.988158] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521c1925-127f-0651-e167-d9ce1b4c928f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.111866] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Releasing lock "refresh_cache-ea4a3f54-388e-451f-9318-78943a169213" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.112121] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.112304] env[62235]: DEBUG nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.112473] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.128854] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.271915] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e9175e-c341-4c78-93b0-183159422a64 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.280030] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0aea71-414f-4ced-844e-723be23d83d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.311213] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8f2590-a041-41c2-b5e2-de6f4f6b12f7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.318017] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dde35c-3896-4367-a789-e27401ebc6a1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.330538] env[62235]: DEBUG nova.compute.provider_tree [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.332333] env[62235]: INFO nova.scheduler.client.report [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Deleted allocations for instance 1c6f4360-b80b-43e3-accc-f7500e626100 [ 754.482769] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271498, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060077} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.486133] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 754.486895] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a8dc5b-a266-44f7-babf-213c3370c211 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.494657] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521c1925-127f-0651-e167-d9ce1b4c928f, 'name': SearchDatastore_Task, 'duration_secs': 0.009156} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.501581] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.501861] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 754.510511] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 754.510792] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80599810-ae16-44ae-b1b0-58873fab36eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.512881] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42a17593-8292-4e95-b164-5431bf1c12b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.534200] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 754.534200] env[62235]: value = "task-1271500" [ 754.534200] env[62235]: _type = "Task" [ 754.534200] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.535469] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 754.535469] env[62235]: value = "task-1271499" [ 754.535469] env[62235]: _type = "Task" [ 754.535469] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.545978] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271500, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.548924] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271499, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.631321] env[62235]: DEBUG nova.network.neutron [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.839525] env[62235]: DEBUG nova.scheduler.client.report [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.844189] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc88c79-93c0-4b54-9930-d84aabc6153f tempest-VolumesAdminNegativeTest-1015263038 tempest-VolumesAdminNegativeTest-1015263038-project-member] Lock "1c6f4360-b80b-43e3-accc-f7500e626100" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 168.128s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.048375] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271499, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46436} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.051614] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 755.051837] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.052098] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271500, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.052307] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75328d7b-2635-4686-9832-d2d101b5c683 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.059437] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 755.059437] env[62235]: value = "task-1271501" [ 755.059437] env[62235]: _type = "Task" [ 755.059437] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.067582] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.134042] env[62235]: INFO nova.compute.manager [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] [instance: ea4a3f54-388e-451f-9318-78943a169213] Took 1.02 seconds to deallocate network for instance. [ 755.350021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.350021] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.350764] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.980s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.352329] env[62235]: INFO nova.compute.claims [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.354933] env[62235]: DEBUG nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.550147] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271500, 'name': ReconfigVM_Task, 'duration_secs': 0.542514} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.550147] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 755.550147] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a8cd3dd-628c-49b9-a895-e6dc82f97d93 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.555585] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 755.555585] env[62235]: value = "task-1271502" [ 755.555585] env[62235]: _type = "Task" [ 755.555585] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.569477] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271502, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.572263] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065612} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.572509] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.573248] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde528ad-f4b0-487e-91bc-96e5c2352c7a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.594791] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.595090] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd27e86a-d52e-49b8-86b0-c32979627380 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.613666] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 755.613666] env[62235]: value = "task-1271503" [ 755.613666] env[62235]: _type = "Task" [ 755.613666] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.621472] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271503, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.864731] env[62235]: DEBUG nova.compute.utils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.866271] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.866705] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 755.921510] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.951121] env[62235]: DEBUG nova.policy [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e5db836b6d94d98af3643a83b5de17a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8fe9bc151d1455eb5c1f728e533e1bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.067019] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271502, 'name': Rename_Task, 'duration_secs': 0.127394} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.067372] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.067639] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0192b7d7-5d25-4423-b2fc-09aa071aa021 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.073707] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 756.073707] env[62235]: value = "task-1271504" [ 756.073707] env[62235]: _type = "Task" [ 756.073707] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.082302] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271504, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.124549] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271503, 'name': ReconfigVM_Task, 'duration_secs': 0.417082} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.124549] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 430056ca-ccba-45c5-adf7-9c068f81f9ec/430056ca-ccba-45c5-adf7-9c068f81f9ec.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.124549] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-745127c0-61e7-4446-86f4-b1a77ed4ccd5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.129959] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 756.129959] env[62235]: value = "task-1271505" [ 756.129959] env[62235]: _type = "Task" [ 756.129959] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.137910] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271505, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.170181] env[62235]: INFO nova.scheduler.client.report [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Deleted allocations for instance ea4a3f54-388e-451f-9318-78943a169213 [ 756.366692] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.388551] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Successfully created port: 47d535bf-048e-404c-825f-0290d9fce21f {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.584420] env[62235]: DEBUG oslo_vmware.api [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271504, 'name': PowerOnVM_Task, 'duration_secs': 0.475669} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.585459] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 756.585786] env[62235]: INFO nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Took 5.32 seconds to spawn the instance on the hypervisor. [ 756.585879] env[62235]: DEBUG nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 756.586745] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e5678c-aad6-4ce1-867b-788e692f86df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.646488] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271505, 'name': Rename_Task, 'duration_secs': 0.163396} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.650074] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.650074] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07ac64bd-917a-42b8-a785-62c35bfbb73e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.657261] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 756.657261] env[62235]: value = "task-1271507" [ 756.657261] env[62235]: _type = "Task" [ 756.657261] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.671578] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.685786] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f818bc1-4051-46c6-aea0-139c7fd8d2f3 tempest-ServersTestBootFromVolume-1118182822 tempest-ServersTestBootFromVolume-1118182822-project-member] Lock "ea4a3f54-388e-451f-9318-78943a169213" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.402s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.728403] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31752a55-784b-4856-8371-97b7e048e5e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.741451] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42deea2-9743-4653-a53f-a1101eaeeb8e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.779746] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a2bd89-61d0-4374-9699-24e002f58d5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.787569] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2404367d-fc90-4b16-8765-497c34a871f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.806555] env[62235]: DEBUG nova.compute.provider_tree [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.108149] env[62235]: INFO nova.compute.manager [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Took 28.66 seconds to build instance. [ 757.168756] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271507, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.188392] env[62235]: DEBUG nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.313442] env[62235]: DEBUG nova.scheduler.client.report [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.363723] env[62235]: DEBUG nova.compute.manager [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Received event network-changed-47d535bf-048e-404c-825f-0290d9fce21f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.363723] env[62235]: DEBUG nova.compute.manager [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Refreshing instance network info cache due to event network-changed-47d535bf-048e-404c-825f-0290d9fce21f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.363723] env[62235]: DEBUG oslo_concurrency.lockutils [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] Acquiring lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.363723] env[62235]: DEBUG oslo_concurrency.lockutils [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] Acquired lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.363723] env[62235]: DEBUG nova.network.neutron [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Refreshing network info cache for port 47d535bf-048e-404c-825f-0290d9fce21f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.381677] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.411447] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.411933] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.412262] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.413088] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.413088] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.413088] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.413088] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.413742] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.414639] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.414639] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.414639] env[62235]: DEBUG nova.virt.hardware [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.415779] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e156a1-ca4a-486b-8b9f-de1a03496c01 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.424412] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c074a3-1782-4e72-932d-edf86c145567 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.540713] env[62235]: ERROR nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. [ 757.540713] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.540713] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.540713] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.540713] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.540713] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.540713] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.540713] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.540713] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.540713] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 757.540713] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.540713] env[62235]: ERROR nova.compute.manager raise self.value [ 757.540713] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.540713] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.540713] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.540713] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.541216] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.541216] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.541216] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. [ 757.541216] env[62235]: ERROR nova.compute.manager [ 757.541216] env[62235]: Traceback (most recent call last): [ 757.541216] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.541216] env[62235]: listener.cb(fileno) [ 757.541216] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.541216] env[62235]: result = function(*args, **kwargs) [ 757.541216] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.541216] env[62235]: return func(*args, **kwargs) [ 757.541216] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.541216] env[62235]: raise e [ 757.541216] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.541216] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 757.541216] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.541216] env[62235]: created_port_ids = self._update_ports_for_instance( [ 757.541216] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.541216] env[62235]: with excutils.save_and_reraise_exception(): [ 757.541216] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.541216] env[62235]: self.force_reraise() [ 757.541216] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.541216] env[62235]: raise self.value [ 757.541216] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.541216] env[62235]: updated_port = self._update_port( [ 757.541216] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.541216] env[62235]: _ensure_no_port_binding_failure(port) [ 757.541216] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.541216] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.542108] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. [ 757.542108] env[62235]: Removing descriptor: 16 [ 757.542108] env[62235]: ERROR nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Traceback (most recent call last): [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] yield resources [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self.driver.spawn(context, instance, image_meta, [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.542108] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] vm_ref = self.build_virtual_machine(instance, [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] for vif in network_info: [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return self._sync_wrapper(fn, *args, **kwargs) [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self.wait() [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self[:] = self._gt.wait() [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return self._exit_event.wait() [ 757.542461] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] result = hub.switch() [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return self.greenlet.switch() [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] result = function(*args, **kwargs) [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return func(*args, **kwargs) [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] raise e [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] nwinfo = self.network_api.allocate_for_instance( [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.542834] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] created_port_ids = self._update_ports_for_instance( [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] with excutils.save_and_reraise_exception(): [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self.force_reraise() [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] raise self.value [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] updated_port = self._update_port( [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] _ensure_no_port_binding_failure(port) [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.543195] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] raise exception.PortBindingFailed(port_id=port['id']) [ 757.543537] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] nova.exception.PortBindingFailed: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. [ 757.543537] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] [ 757.543537] env[62235]: INFO nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Terminating instance [ 757.546268] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Acquiring lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.615514] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3bb59af1-479e-48bc-b83f-65675722e788 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.235s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.671591] env[62235]: DEBUG oslo_vmware.api [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271507, 'name': PowerOnVM_Task, 'duration_secs': 0.838168} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.672088] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 757.672421] env[62235]: DEBUG nova.compute.manager [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.673570] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8fad7c-bdaf-4745-8127-ecbe2191f717 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.723022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.819222] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.819722] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.822305] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.316s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.894124] env[62235]: DEBUG nova.network.neutron [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.989681] env[62235]: DEBUG nova.network.neutron [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.120670] env[62235]: DEBUG nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.208175] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.265019] env[62235]: INFO nova.compute.manager [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Rebuilding instance [ 758.331761] env[62235]: DEBUG nova.compute.utils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.340859] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 758.341988] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 758.344659] env[62235]: DEBUG nova.compute.manager [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 758.349263] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba51c51d-874d-4638-a433-0529d72a3d8b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.408622] env[62235]: DEBUG nova.policy [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19727a6117df4d298b926f707483a8e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2171f5abf1ef4e1987b9a77d132215bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.495324] env[62235]: DEBUG oslo_concurrency.lockutils [req-a032a79a-e157-49c0-b192-2086043f7fe3 req-2237959c-fa9e-4682-8cb0-36db462524e6 service nova] Releasing lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.496254] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Acquired lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.496254] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.640218] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.645694] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fd81a9-c1d0-4964-88de-5405480382b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.653681] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f221b4-f9b0-4409-80ee-6839cf0bc6b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.688307] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd19d4bb-c6aa-493d-9773-04e51f1fe570 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.696156] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974d44ea-bf2a-4188-9f49-9f67ec82cea1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.709589] env[62235]: DEBUG nova.compute.provider_tree [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.739343] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Successfully created port: 9cad198b-7619-4450-8e94-47b42ba1d8a6 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.845261] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.870714] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.871046] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8996a71-2c70-41ec-9319-c81914173cbb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.885786] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 758.885786] env[62235]: value = "task-1271508" [ 758.885786] env[62235]: _type = "Task" [ 758.885786] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.897880] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.037729] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.212903] env[62235]: DEBUG nova.scheduler.client.report [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.297664] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.361313] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "430056ca-ccba-45c5-adf7-9c068f81f9ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.362033] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "430056ca-ccba-45c5-adf7-9c068f81f9ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.362033] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "430056ca-ccba-45c5-adf7-9c068f81f9ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.362284] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "430056ca-ccba-45c5-adf7-9c068f81f9ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.362558] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "430056ca-ccba-45c5-adf7-9c068f81f9ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.367500] env[62235]: INFO nova.compute.manager [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Terminating instance [ 759.370090] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "refresh_cache-430056ca-ccba-45c5-adf7-9c068f81f9ec" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.370335] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquired lock "refresh_cache-430056ca-ccba-45c5-adf7-9c068f81f9ec" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.370715] env[62235]: DEBUG nova.network.neutron [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.402065] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271508, 'name': PowerOffVM_Task, 'duration_secs': 0.240629} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.402484] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 759.402890] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.404860] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a84c8f-83ee-4109-9fed-b4e1c7599a07 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.416565] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 759.416565] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59acdc9c-91dc-4f48-98c1-f4a79aba513d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.427181] env[62235]: DEBUG nova.compute.manager [req-0611a398-db91-4e3d-8517-b658c5b80128 req-21314989-22a5-436f-b9e6-986fc094bb20 service nova] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Received event network-vif-deleted-47d535bf-048e-404c-825f-0290d9fce21f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.453410] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 759.453711] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 759.453906] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Deleting the datastore file [datastore2] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.454188] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4980a62-fc6a-494f-83be-298e4d8a6071 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.461581] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 759.461581] env[62235]: value = "task-1271510" [ 759.461581] env[62235]: _type = "Task" [ 759.461581] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.471475] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271510, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.719763] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.720445] env[62235]: ERROR nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Traceback (most recent call last): [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self.driver.spawn(context, instance, image_meta, [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] vm_ref = self.build_virtual_machine(instance, [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.720445] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] for vif in network_info: [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return self._sync_wrapper(fn, *args, **kwargs) [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self.wait() [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self[:] = self._gt.wait() [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return self._exit_event.wait() [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] result = hub.switch() [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 759.720845] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return self.greenlet.switch() [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] result = function(*args, **kwargs) [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] return func(*args, **kwargs) [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] raise e [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] nwinfo = self.network_api.allocate_for_instance( [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] created_port_ids = self._update_ports_for_instance( [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] with excutils.save_and_reraise_exception(): [ 759.721216] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] self.force_reraise() [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] raise self.value [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] updated_port = self._update_port( [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] _ensure_no_port_binding_failure(port) [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] raise exception.PortBindingFailed(port_id=port['id']) [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] nova.exception.PortBindingFailed: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. [ 759.721776] env[62235]: ERROR nova.compute.manager [instance: 3b4167dd-4535-4482-9452-5548e07dc581] [ 759.722146] env[62235]: DEBUG nova.compute.utils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.722777] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.848s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.726808] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Build of instance 3b4167dd-4535-4482-9452-5548e07dc581 was re-scheduled: Binding failed for port cbee762a-ec5f-4935-b053-b9b3bc7eae0b, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.727932] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 759.728267] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Acquiring lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.728879] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Acquired lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.729283] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.800273] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Releasing lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.800707] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.800973] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.801433] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3fb36044-a286-4f12-abf6-11cb937338c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.816024] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d6f98c-41e1-46dd-80b9-0adf875ac959 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.841218] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6671163c-c829-4046-97e5-e85fe5e223dd could not be found. [ 759.841451] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.841634] env[62235]: INFO nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 759.841911] env[62235]: DEBUG oslo.service.loopingcall [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.842163] env[62235]: DEBUG nova.compute.manager [-] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.842249] env[62235]: DEBUG nova.network.neutron [-] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.859765] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.868829] env[62235]: DEBUG nova.network.neutron [-] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.894062] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.894347] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.894505] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.894707] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.894881] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.895049] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.896402] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.896550] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.896775] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.896985] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.897200] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.898541] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346c7e6e-4994-4fd1-b9ce-35c71f22738e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.906354] env[62235]: DEBUG nova.network.neutron [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.910418] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73409680-2b9f-4ff6-b6cf-39694bc6957b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.978199] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271510, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.08852} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.978638] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.978959] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.979390] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.008591] env[62235]: DEBUG nova.network.neutron [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.265010] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.372656] env[62235]: DEBUG nova.network.neutron [-] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.427955] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.512086] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Releasing lock "refresh_cache-430056ca-ccba-45c5-adf7-9c068f81f9ec" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.512717] env[62235]: DEBUG nova.compute.manager [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.514954] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.517887] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b0935f-f458-4271-81c6-4ad34f59c1be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.529940] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 760.530628] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60eb12ef-8174-4569-976a-d1ebc36b1485 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.539064] env[62235]: DEBUG oslo_vmware.api [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 760.539064] env[62235]: value = "task-1271511" [ 760.539064] env[62235]: _type = "Task" [ 760.539064] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.550296] env[62235]: DEBUG oslo_vmware.api [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.651502] env[62235]: ERROR nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. [ 760.651502] env[62235]: ERROR nova.compute.manager Traceback (most recent call last): [ 760.651502] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.651502] env[62235]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 760.651502] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.651502] env[62235]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 760.651502] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.651502] env[62235]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 760.651502] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.651502] env[62235]: ERROR nova.compute.manager self.force_reraise() [ 760.651502] env[62235]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.651502] env[62235]: ERROR nova.compute.manager raise self.value [ 760.651502] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.651502] env[62235]: ERROR nova.compute.manager updated_port = self._update_port( [ 760.651502] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.651502] env[62235]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 760.652085] env[62235]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.652085] env[62235]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 760.652085] env[62235]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. [ 760.652085] env[62235]: ERROR nova.compute.manager [ 760.652217] env[62235]: Traceback (most recent call last): [ 760.652284] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 760.652284] env[62235]: listener.cb(fileno) [ 760.652284] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.652284] env[62235]: result = function(*args, **kwargs) [ 760.652412] env[62235]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.652412] env[62235]: return func(*args, **kwargs) [ 760.652412] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.652412] env[62235]: raise e [ 760.652412] env[62235]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.652412] env[62235]: nwinfo = self.network_api.allocate_for_instance( [ 760.652412] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.652412] env[62235]: created_port_ids = self._update_ports_for_instance( [ 760.652412] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.652412] env[62235]: with excutils.save_and_reraise_exception(): [ 760.652412] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.652412] env[62235]: self.force_reraise() [ 760.652412] env[62235]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.652412] env[62235]: raise self.value [ 760.652412] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.652412] env[62235]: updated_port = self._update_port( [ 760.652412] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.652412] env[62235]: _ensure_no_port_binding_failure(port) [ 760.652412] env[62235]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.652412] env[62235]: raise exception.PortBindingFailed(port_id=port['id']) [ 760.652412] env[62235]: nova.exception.PortBindingFailed: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. [ 760.652412] env[62235]: Removing descriptor: 16 [ 760.654502] env[62235]: ERROR nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Traceback (most recent call last): [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] yield resources [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self.driver.spawn(context, instance, image_meta, [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] vm_ref = self.build_virtual_machine(instance, [ 760.654502] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] for vif in network_info: [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return self._sync_wrapper(fn, *args, **kwargs) [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self.wait() [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self[:] = self._gt.wait() [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return self._exit_event.wait() [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.654886] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] result = hub.switch() [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return self.greenlet.switch() [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] result = function(*args, **kwargs) [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return func(*args, **kwargs) [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] raise e [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] nwinfo = self.network_api.allocate_for_instance( [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] created_port_ids = self._update_ports_for_instance( [ 760.656184] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] with excutils.save_and_reraise_exception(): [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self.force_reraise() [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] raise self.value [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] updated_port = self._update_port( [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] _ensure_no_port_binding_failure(port) [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] raise exception.PortBindingFailed(port_id=port['id']) [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] nova.exception.PortBindingFailed: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. [ 760.656585] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] [ 760.657138] env[62235]: INFO nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Terminating instance [ 760.659956] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6c4a82-d540-4a82-8189-da226fe58c41 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.665348] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.665758] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.666145] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.673919] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8837987-40d7-4852-bbd2-1b176b98f32c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.711847] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c42dbc-42bf-42d3-9250-72b8a120108e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.719630] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f827927c-16e7-481c-8aeb-f082df5dc9fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.734172] env[62235]: DEBUG nova.compute.provider_tree [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.874578] env[62235]: INFO nova.compute.manager [-] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Took 1.03 seconds to deallocate network for instance. [ 760.878787] env[62235]: DEBUG nova.compute.claims [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.878787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.933753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Releasing lock "refresh_cache-3b4167dd-4535-4482-9452-5548e07dc581" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.934029] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 760.934360] env[62235]: DEBUG nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.934438] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.954703] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.020752] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.021865] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.021865] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.021865] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.021865] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.021865] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.021865] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.022187] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.022187] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.022292] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.022460] env[62235]: DEBUG nova.virt.hardware [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.023352] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384c39b1-431c-4c4c-851b-a2c929fd3090 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.032312] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e683b109-4e40-43b2-bf3e-6d21d63036e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.045397] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.051334] env[62235]: DEBUG oslo.service.loopingcall [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.054541] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.055046] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4538a325-9166-4b46-aa83-92640833c9f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.071242] env[62235]: DEBUG oslo_vmware.api [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271511, 'name': PowerOffVM_Task, 'duration_secs': 0.226181} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.072494] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 761.072676] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 761.072904] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.072904] env[62235]: value = "task-1271512" [ 761.072904] env[62235]: _type = "Task" [ 761.072904] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.073098] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-161644b5-cfa9-4072-a4df-d8978e966329 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.082185] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271512, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.099154] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 761.099437] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 761.099675] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Deleting the datastore file [datastore2] 430056ca-ccba-45c5-adf7-9c068f81f9ec {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 761.100390] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21f5add0-b625-4be1-8700-fe5b5e9bffb2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.111009] env[62235]: DEBUG oslo_vmware.api [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for the task: (returnval){ [ 761.111009] env[62235]: value = "task-1271514" [ 761.111009] env[62235]: _type = "Task" [ 761.111009] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.119886] env[62235]: DEBUG oslo_vmware.api [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271514, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.194629] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.237844] env[62235]: DEBUG nova.scheduler.client.report [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.329682] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.458830] env[62235]: DEBUG nova.network.neutron [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.467599] env[62235]: DEBUG nova.compute.manager [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Received event network-changed-9cad198b-7619-4450-8e94-47b42ba1d8a6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.467599] env[62235]: DEBUG nova.compute.manager [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Refreshing instance network info cache due to event network-changed-9cad198b-7619-4450-8e94-47b42ba1d8a6. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.467599] env[62235]: DEBUG oslo_concurrency.lockutils [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] Acquiring lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.585377] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271512, 'name': CreateVM_Task, 'duration_secs': 0.324333} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.585545] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.585971] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.586157] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.586482] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.586725] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ea2018a-422c-4a26-9d69-0e25913bdfbe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.591366] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 761.591366] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5235f5a7-ee8d-f1a0-0862-7e7e2f366db1" [ 761.591366] env[62235]: _type = "Task" [ 761.591366] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.599468] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5235f5a7-ee8d-f1a0-0862-7e7e2f366db1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.619950] env[62235]: DEBUG oslo_vmware.api [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Task: {'id': task-1271514, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145051} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.620698] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.620698] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 761.620848] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.621013] env[62235]: INFO nova.compute.manager [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Took 1.11 seconds to destroy the instance on the hypervisor. [ 761.621261] env[62235]: DEBUG oslo.service.loopingcall [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.621735] env[62235]: DEBUG nova.compute.manager [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.621840] env[62235]: DEBUG nova.network.neutron [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.643959] env[62235]: DEBUG nova.network.neutron [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.749177] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.026s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.749806] env[62235]: ERROR nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] Traceback (most recent call last): [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self.driver.spawn(context, instance, image_meta, [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] vm_ref = self.build_virtual_machine(instance, [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.749806] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] for vif in network_info: [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return self._sync_wrapper(fn, *args, **kwargs) [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self.wait() [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self[:] = self._gt.wait() [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return self._exit_event.wait() [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] result = hub.switch() [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.750325] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return self.greenlet.switch() [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] result = function(*args, **kwargs) [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] return func(*args, **kwargs) [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] raise e [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] nwinfo = self.network_api.allocate_for_instance( [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] created_port_ids = self._update_ports_for_instance( [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] with excutils.save_and_reraise_exception(): [ 761.751009] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] self.force_reraise() [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] raise self.value [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] updated_port = self._update_port( [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] _ensure_no_port_binding_failure(port) [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] raise exception.PortBindingFailed(port_id=port['id']) [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] nova.exception.PortBindingFailed: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. [ 761.751563] env[62235]: ERROR nova.compute.manager [instance: 86e61430-ce9a-436e-9331-6276604610e3] [ 761.752221] env[62235]: DEBUG nova.compute.utils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.752221] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.157s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.755059] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Build of instance 86e61430-ce9a-436e-9331-6276604610e3 was re-scheduled: Binding failed for port 4216d720-b7ac-44ef-8387-21fe8582badc, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.755491] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.755709] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.755856] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.756058] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.834236] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.834905] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.835234] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 761.835684] env[62235]: DEBUG oslo_concurrency.lockutils [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] Acquired lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.835973] env[62235]: DEBUG nova.network.neutron [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Refreshing network info cache for port 9cad198b-7619-4450-8e94-47b42ba1d8a6 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 761.837453] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbff4e7d-80fb-4412-ba5a-db0c4d8ff2e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.852847] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fec6b8b-e202-4d33-90d2-64c1482a02fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.925141] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3e7a81ae-2163-4562-b483-dc639c602fa0 could not be found. [ 761.925141] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.925141] env[62235]: INFO nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Took 0.05 seconds to destroy the instance on the hypervisor. [ 761.925141] env[62235]: DEBUG oslo.service.loopingcall [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.925141] env[62235]: DEBUG nova.compute.manager [-] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.925141] env[62235]: DEBUG nova.network.neutron [-] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.929801] env[62235]: DEBUG nova.network.neutron [-] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.961462] env[62235]: INFO nova.compute.manager [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] [instance: 3b4167dd-4535-4482-9452-5548e07dc581] Took 1.03 seconds to deallocate network for instance. [ 762.101511] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5235f5a7-ee8d-f1a0-0862-7e7e2f366db1, 'name': SearchDatastore_Task, 'duration_secs': 0.011106} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.101818] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.102063] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.102300] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.102444] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.102627] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.102873] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa8d7ee4-8445-4b43-a47c-7850697b10e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.111224] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.111528] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.112136] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cee9a420-5f7d-4b24-9a2c-e968353222fa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.117534] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 762.117534] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52789a82-edb8-b225-274f-63170bb9cfce" [ 762.117534] env[62235]: _type = "Task" [ 762.117534] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.125387] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52789a82-edb8-b225-274f-63170bb9cfce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.144842] env[62235]: DEBUG nova.network.neutron [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.282917] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.368935] env[62235]: DEBUG nova.network.neutron [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.410279] env[62235]: DEBUG nova.network.neutron [-] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.463033] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.565110] env[62235]: DEBUG nova.network.neutron [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.628234] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52789a82-edb8-b225-274f-63170bb9cfce, 'name': SearchDatastore_Task, 'duration_secs': 0.008206} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.631976] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-271d14b9-1bb4-487d-b9ee-45b6107e61be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.637486] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 762.637486] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52715402-f67e-49e4-d9dd-4515c94cfbda" [ 762.637486] env[62235]: _type = "Task" [ 762.637486] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.648589] env[62235]: INFO nova.compute.manager [-] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Took 1.03 seconds to deallocate network for instance. [ 762.648886] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52715402-f67e-49e4-d9dd-4515c94cfbda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.662846] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d68b0b-7b02-456c-9589-3785ce92dd1d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.669814] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e46dcd0-d052-4e4f-bd14-65422fef2aa4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.702521] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7224970b-acc0-4c15-babc-72ac0cab2deb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.710466] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2383eef5-9092-461d-ae58-4a8bdda2ab80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.726387] env[62235]: DEBUG nova.compute.provider_tree [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.915019] env[62235]: INFO nova.compute.manager [-] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Took 1.03 seconds to deallocate network for instance. [ 762.917180] env[62235]: DEBUG nova.compute.claims [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Aborting claim: {{(pid=62235) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 762.917360] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.966176] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-86e61430-ce9a-436e-9331-6276604610e3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.966445] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.966634] env[62235]: DEBUG nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.966801] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.986247] env[62235]: INFO nova.scheduler.client.report [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Deleted allocations for instance 3b4167dd-4535-4482-9452-5548e07dc581 [ 762.998125] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.070041] env[62235]: DEBUG oslo_concurrency.lockutils [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] Releasing lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.070309] env[62235]: DEBUG nova.compute.manager [req-576043ae-0ee8-4792-ac3d-a8519f00b1ae req-e08691fe-4944-4173-ac24-4815d5abf4fc service nova] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Received event network-vif-deleted-9cad198b-7619-4450-8e94-47b42ba1d8a6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.148090] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52715402-f67e-49e4-d9dd-4515c94cfbda, 'name': SearchDatastore_Task, 'duration_secs': 0.00863} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.149352] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.149352] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 763.149352] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89532ef8-10bb-46a5-a24c-d05d474d1d27 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.155115] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.156984] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 763.156984] env[62235]: value = "task-1271515" [ 763.156984] env[62235]: _type = "Task" [ 763.156984] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.163950] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271515, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.229785] env[62235]: DEBUG nova.scheduler.client.report [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.500716] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bfa42c-0d3d-4661-bbf9-6b5ec54b8999 tempest-ServersNegativeTestMultiTenantJSON-556835672 tempest-ServersNegativeTestMultiTenantJSON-556835672-project-member] Lock "3b4167dd-4535-4482-9452-5548e07dc581" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.792s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.501161] env[62235]: DEBUG nova.network.neutron [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.667821] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271515, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457219} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.668318] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 763.672757] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.672757] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb408867-a8f6-4479-9751-4b7648b8fc28 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.677129] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 763.677129] env[62235]: value = "task-1271516" [ 763.677129] env[62235]: _type = "Task" [ 763.677129] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.686320] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271516, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.735697] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.736397] env[62235]: ERROR nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Traceback (most recent call last): [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self.driver.spawn(context, instance, image_meta, [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] vm_ref = self.build_virtual_machine(instance, [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.736397] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] for vif in network_info: [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] return self._sync_wrapper(fn, *args, **kwargs) [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self.wait() [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self[:] = self._gt.wait() [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] return self._exit_event.wait() [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] current.throw(*self._exc) [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.736624] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] result = function(*args, **kwargs) [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] return func(*args, **kwargs) [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] raise e [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] nwinfo = self.network_api.allocate_for_instance( [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] created_port_ids = self._update_ports_for_instance( [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] with excutils.save_and_reraise_exception(): [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] self.force_reraise() [ 763.736869] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] raise self.value [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] updated_port = self._update_port( [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] _ensure_no_port_binding_failure(port) [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] raise exception.PortBindingFailed(port_id=port['id']) [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] nova.exception.PortBindingFailed: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. [ 763.737134] env[62235]: ERROR nova.compute.manager [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] [ 763.737325] env[62235]: DEBUG nova.compute.utils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 763.738715] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.380s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.740379] env[62235]: INFO nova.compute.claims [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.743052] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Build of instance c62172db-7e1c-4fd1-b8b9-a8daf0a71f91 was re-scheduled: Binding failed for port 524e4b31-e355-46e2-b753-4d5e24d4151b, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 763.743485] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 763.743711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.743884] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquired lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.744063] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.005119] env[62235]: INFO nova.compute.manager [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 86e61430-ce9a-436e-9331-6276604610e3] Took 1.04 seconds to deallocate network for instance. [ 764.008066] env[62235]: DEBUG nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.188148] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271516, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067795} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.188444] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.189232] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359aca22-f9f8-4a97-917a-9e812c1646f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.214600] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.214978] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e03d219-00a1-4787-9a47-177a696b740f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.238740] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 764.238740] env[62235]: value = "task-1271517" [ 764.238740] env[62235]: _type = "Task" [ 764.238740] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.256019] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271517, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.270879] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.360283] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.540778] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.751599] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271517, 'name': ReconfigVM_Task, 'duration_secs': 0.280754} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.752049] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba/8559690a-bdd3-4ec3-98e1-dfc78c6db8ba.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 764.752675] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28235a76-5b69-4248-a2d2-fc5f19aefcb7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.761147] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 764.761147] env[62235]: value = "task-1271518" [ 764.761147] env[62235]: _type = "Task" [ 764.761147] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.769576] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271518, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.867304] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Releasing lock "refresh_cache-c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.867489] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 764.867652] env[62235]: DEBUG nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.867825] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.898371] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.038056] env[62235]: INFO nova.scheduler.client.report [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted allocations for instance 86e61430-ce9a-436e-9331-6276604610e3 [ 765.114962] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9ead95-e0a5-4f97-91ea-13c703a935d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.123018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18409bc-f1ec-4ce8-8b9d-5788fe5838bb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.154162] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a97848-5264-4aac-a427-b1a307eec580 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.161208] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a652c1-a789-48c9-8431-ac7335bc8352 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.176395] env[62235]: DEBUG nova.compute.provider_tree [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.267200] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "80a803d1-c034-42b6-a15d-9fc3893953ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.267200] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.281216] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271518, 'name': Rename_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.407832] env[62235]: DEBUG nova.network.neutron [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.550495] env[62235]: DEBUG oslo_concurrency.lockutils [None req-009fd1a0-251d-4548-9fbb-d65c72a81274 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "86e61430-ce9a-436e-9331-6276604610e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.726s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.679864] env[62235]: DEBUG nova.scheduler.client.report [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.777032] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271518, 'name': Rename_Task, 'duration_secs': 0.853606} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.777032] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 765.777032] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-173b6874-dc3c-436f-897d-59d73963b30c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.781690] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 765.781690] env[62235]: value = "task-1271519" [ 765.781690] env[62235]: _type = "Task" [ 765.781690] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.789185] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271519, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.914819] env[62235]: INFO nova.compute.manager [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: c62172db-7e1c-4fd1-b8b9-a8daf0a71f91] Took 1.05 seconds to deallocate network for instance. [ 766.053937] env[62235]: DEBUG nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.186940] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.187467] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.190100] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.354s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.191832] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.192043] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 766.192728] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.168s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.197495] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77486ff-c5f9-4931-856a-d69fadb72c33 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.205789] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad56cd3a-c09f-4f60-8dd4-b30bd48444cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.222552] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc2ed84-c644-48ad-88b3-297bcf372f85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.233101] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35536185-1f43-4722-a031-6855c4577ee5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.267494] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181540MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 766.267916] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.290622] env[62235]: DEBUG oslo_vmware.api [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271519, 'name': PowerOnVM_Task, 'duration_secs': 0.415435} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.290888] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 766.291102] env[62235]: DEBUG nova.compute.manager [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 766.291874] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b949bf4-c2fa-429a-8ce6-77da35c73540 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.578675] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.701871] env[62235]: DEBUG nova.compute.utils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.703805] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.704090] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 766.777944] env[62235]: DEBUG nova.policy [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19727a6117df4d298b926f707483a8e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2171f5abf1ef4e1987b9a77d132215bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 766.808326] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.955977] env[62235]: INFO nova.scheduler.client.report [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Deleted allocations for instance c62172db-7e1c-4fd1-b8b9-a8daf0a71f91 [ 767.058498] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bc6c4a-7f2c-44e8-8ded-77da4f6888f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.068017] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0a164b-5999-4571-9f0e-0097acdcea63 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.104017] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c3f5ef-69e3-4378-832a-c87123f00a2e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.114389] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9e6b9e-cc90-4a67-a647-18e9e5e3f4e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.121451] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.121575] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.121773] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.121952] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.122134] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.124325] env[62235]: INFO nova.compute.manager [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Terminating instance [ 767.126013] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "refresh_cache-8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.126182] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquired lock "refresh_cache-8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.126365] env[62235]: DEBUG nova.network.neutron [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.136290] env[62235]: DEBUG nova.compute.provider_tree [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.208357] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.217254] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Successfully created port: da476e21-90c9-4f68-926c-883707c320ca {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.255571] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.256335] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.466449] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bb5feeee-1a81-45d1-ac85-7245d20ac2d0 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "c62172db-7e1c-4fd1-b8b9-a8daf0a71f91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.231s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.640669] env[62235]: DEBUG nova.scheduler.client.report [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.658015] env[62235]: DEBUG nova.network.neutron [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.734179] env[62235]: DEBUG nova.network.neutron [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.969263] env[62235]: DEBUG nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.145518] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.146616] env[62235]: ERROR nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Traceback (most recent call last): [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self.driver.spawn(context, instance, image_meta, [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] vm_ref = self.build_virtual_machine(instance, [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.146616] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] for vif in network_info: [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return self._sync_wrapper(fn, *args, **kwargs) [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self.wait() [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self[:] = self._gt.wait() [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return self._exit_event.wait() [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] result = hub.switch() [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 768.146913] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return self.greenlet.switch() [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] result = function(*args, **kwargs) [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] return func(*args, **kwargs) [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] raise e [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] nwinfo = self.network_api.allocate_for_instance( [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] created_port_ids = self._update_ports_for_instance( [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] with excutils.save_and_reraise_exception(): [ 768.147283] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] self.force_reraise() [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] raise self.value [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] updated_port = self._update_port( [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] _ensure_no_port_binding_failure(port) [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] raise exception.PortBindingFailed(port_id=port['id']) [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] nova.exception.PortBindingFailed: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. [ 768.147603] env[62235]: ERROR nova.compute.manager [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] [ 768.148389] env[62235]: DEBUG nova.compute.utils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 768.149854] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.229s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.153107] env[62235]: INFO nova.compute.claims [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.155870] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Build of instance 7e6ebba9-58e8-4971-b9fb-8365a61d5505 was re-scheduled: Binding failed for port 932dd0ad-9494-4ebf-b0c0-ff87b847e67d, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 768.156346] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 768.156571] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.156709] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquired lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.156861] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.220199] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.237256] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Releasing lock "refresh_cache-8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.237256] env[62235]: DEBUG nova.compute.manager [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.237256] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.237715] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69c619a-e016-46f5-9e90-3a524f44c5a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.247307] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.247550] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bcc6e17b-8b3f-48be-b777-c51e4575238f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.250998] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.251238] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.251392] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.251571] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.251711] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.251852] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.252131] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.252258] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.252422] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.252577] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.252742] env[62235]: DEBUG nova.virt.hardware [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.253503] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e07f196-aa3f-48e3-9177-62603de0d9c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.264165] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b8149-f4a4-41f0-b7e2-af0bb455a5bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.268698] env[62235]: DEBUG oslo_vmware.api [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 768.268698] env[62235]: value = "task-1271520" [ 768.268698] env[62235]: _type = "Task" [ 768.268698] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.285936] env[62235]: DEBUG oslo_vmware.api [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.493059] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.688073] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.780646] env[62235]: DEBUG oslo_vmware.api [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271520, 'name': PowerOffVM_Task, 'duration_secs': 0.147461} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.780930] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.781302] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.781572] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6035f38d-81c2-4d40-b2d7-6a2704fd473f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.807859] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.808094] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.808280] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Deleting the datastore file [datastore1] 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.808531] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-066aabc7-8d18-4ecf-be0f-2bfddd84d7f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.815490] env[62235]: DEBUG oslo_vmware.api [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for the task: (returnval){ [ 768.815490] env[62235]: value = "task-1271522" [ 768.815490] env[62235]: _type = "Task" [ 768.815490] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.823240] env[62235]: DEBUG oslo_vmware.api [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.061320] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.078301] env[62235]: DEBUG nova.compute.manager [req-9efbb9e9-e260-45c7-918b-02ef823c9091 req-5cd88b45-3464-42a2-8aab-acbb49bcb7a6 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Received event network-vif-plugged-da476e21-90c9-4f68-926c-883707c320ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.078521] env[62235]: DEBUG oslo_concurrency.lockutils [req-9efbb9e9-e260-45c7-918b-02ef823c9091 req-5cd88b45-3464-42a2-8aab-acbb49bcb7a6 service nova] Acquiring lock "dd7268d1-1da5-4e5f-9d45-816df147b564-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.078724] env[62235]: DEBUG oslo_concurrency.lockutils [req-9efbb9e9-e260-45c7-918b-02ef823c9091 req-5cd88b45-3464-42a2-8aab-acbb49bcb7a6 service nova] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.078887] env[62235]: DEBUG oslo_concurrency.lockutils [req-9efbb9e9-e260-45c7-918b-02ef823c9091 req-5cd88b45-3464-42a2-8aab-acbb49bcb7a6 service nova] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.079060] env[62235]: DEBUG nova.compute.manager [req-9efbb9e9-e260-45c7-918b-02ef823c9091 req-5cd88b45-3464-42a2-8aab-acbb49bcb7a6 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] No waiting events found dispatching network-vif-plugged-da476e21-90c9-4f68-926c-883707c320ca {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 769.079223] env[62235]: WARNING nova.compute.manager [req-9efbb9e9-e260-45c7-918b-02ef823c9091 req-5cd88b45-3464-42a2-8aab-acbb49bcb7a6 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Received unexpected event network-vif-plugged-da476e21-90c9-4f68-926c-883707c320ca for instance with vm_state building and task_state spawning. [ 769.208143] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Successfully updated port: da476e21-90c9-4f68-926c-883707c320ca {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 769.326631] env[62235]: DEBUG oslo_vmware.api [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Task: {'id': task-1271522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093669} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.329088] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.329468] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.329671] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.329846] env[62235]: INFO nova.compute.manager [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Took 1.09 seconds to destroy the instance on the hypervisor. [ 769.330106] env[62235]: DEBUG oslo.service.loopingcall [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.330935] env[62235]: DEBUG nova.compute.manager [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.330935] env[62235]: DEBUG nova.network.neutron [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.354049] env[62235]: DEBUG nova.network.neutron [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.447262] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96577b7c-808b-421d-97d3-4d30f9dc015e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.454635] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e5467c-a9f6-4030-9334-d19c8c96812a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.485258] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd771b9-025a-45c4-b4eb-e6761f9ef94e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.492626] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d491dd-1fc2-4c8f-ac02-e9627979e3e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.506420] env[62235]: DEBUG nova.compute.provider_tree [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.564370] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Releasing lock "refresh_cache-7e6ebba9-58e8-4971-b9fb-8365a61d5505" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.564573] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 769.564750] env[62235]: DEBUG nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.564916] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.578160] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.712182] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "refresh_cache-dd7268d1-1da5-4e5f-9d45-816df147b564" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.712347] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "refresh_cache-dd7268d1-1da5-4e5f-9d45-816df147b564" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.712501] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.856065] env[62235]: DEBUG nova.network.neutron [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.009868] env[62235]: DEBUG nova.scheduler.client.report [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.081107] env[62235]: DEBUG nova.network.neutron [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.250236] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.358336] env[62235]: INFO nova.compute.manager [-] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Took 1.03 seconds to deallocate network for instance. [ 770.485838] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Updating instance_info_cache with network_info: [{"id": "da476e21-90c9-4f68-926c-883707c320ca", "address": "fa:16:3e:22:40:24", "network": {"id": "01e8e9f0-a88d-4080-a81e-8f5597095aef", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-635657508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2171f5abf1ef4e1987b9a77d132215bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda476e21-90", "ovs_interfaceid": "da476e21-90c9-4f68-926c-883707c320ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.515896] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.516501] env[62235]: DEBUG nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.519699] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.800s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.521328] env[62235]: INFO nova.compute.claims [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.587227] env[62235]: INFO nova.compute.manager [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: 7e6ebba9-58e8-4971-b9fb-8365a61d5505] Took 1.02 seconds to deallocate network for instance. [ 770.869620] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.989847] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "refresh_cache-dd7268d1-1da5-4e5f-9d45-816df147b564" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.990419] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Instance network_info: |[{"id": "da476e21-90c9-4f68-926c-883707c320ca", "address": "fa:16:3e:22:40:24", "network": {"id": "01e8e9f0-a88d-4080-a81e-8f5597095aef", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-635657508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2171f5abf1ef4e1987b9a77d132215bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda476e21-90", "ovs_interfaceid": "da476e21-90c9-4f68-926c-883707c320ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 770.990791] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:40:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '52358fcc-0d9f-45dd-8c75-db533fd992c3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da476e21-90c9-4f68-926c-883707c320ca', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.998975] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Creating folder: Project (2171f5abf1ef4e1987b9a77d132215bc). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 770.999295] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-369cf788-d210-43ea-a781-b4076f2cb6b3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.012360] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Created folder: Project (2171f5abf1ef4e1987b9a77d132215bc) in parent group-v273362. [ 771.012562] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Creating folder: Instances. Parent ref: group-v273390. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 771.013069] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acd6a051-78c4-441f-a517-dcf51ad93132 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.021835] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Created folder: Instances in parent group-v273390. [ 771.021835] env[62235]: DEBUG oslo.service.loopingcall [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.021937] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 771.022116] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e05f86a-55cb-4465-8af8-e73fa6b293c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.038267] env[62235]: DEBUG nova.compute.utils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.041499] env[62235]: DEBUG nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.041673] env[62235]: DEBUG nova.network.neutron [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 771.050525] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.050525] env[62235]: value = "task-1271525" [ 771.050525] env[62235]: _type = "Task" [ 771.050525] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.059897] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271525, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.097749] env[62235]: DEBUG nova.policy [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17a378a1f255481e9ae462eeeed5a82d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e544511a0a5437faa3d65a8205ed3bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.103049] env[62235]: DEBUG nova.compute.manager [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Received event network-changed-da476e21-90c9-4f68-926c-883707c320ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.103311] env[62235]: DEBUG nova.compute.manager [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Refreshing instance network info cache due to event network-changed-da476e21-90c9-4f68-926c-883707c320ca. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.103554] env[62235]: DEBUG oslo_concurrency.lockutils [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] Acquiring lock "refresh_cache-dd7268d1-1da5-4e5f-9d45-816df147b564" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.103704] env[62235]: DEBUG oslo_concurrency.lockutils [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] Acquired lock "refresh_cache-dd7268d1-1da5-4e5f-9d45-816df147b564" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.103886] env[62235]: DEBUG nova.network.neutron [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Refreshing network info cache for port da476e21-90c9-4f68-926c-883707c320ca {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 771.367437] env[62235]: DEBUG nova.network.neutron [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Successfully created port: ca4a4847-2fe0-4db9-b4af-43c3870d60a5 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.542330] env[62235]: DEBUG nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.565245] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271525, 'name': CreateVM_Task, 'duration_secs': 0.280736} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.565245] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 771.585787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.586586] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.586586] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.587046] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ade717ce-6a07-452d-8c36-baecc5f5c98a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.593348] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 771.593348] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ebdde1-17e0-60e7-39f2-fc9c52b7bd23" [ 771.593348] env[62235]: _type = "Task" [ 771.593348] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.612378] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ebdde1-17e0-60e7-39f2-fc9c52b7bd23, 'name': SearchDatastore_Task, 'duration_secs': 0.008192} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.613120] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.613120] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 771.613226] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.613279] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.613456] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 771.613965] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f8ea859-071f-4c74-937d-05d4069b27a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.623084] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.623378] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.627588] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 771.628258] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 771.628545] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10504012-c189-4c4c-bcdf-8b48ea1c165a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.638042] env[62235]: INFO nova.scheduler.client.report [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Deleted allocations for instance 7e6ebba9-58e8-4971-b9fb-8365a61d5505 [ 771.645519] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 771.645519] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5291cc19-6e60-2a6d-3e56-8ceee060eeef" [ 771.645519] env[62235]: _type = "Task" [ 771.645519] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.658310] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5291cc19-6e60-2a6d-3e56-8ceee060eeef, 'name': SearchDatastore_Task, 'duration_secs': 0.007329} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.660925] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e25c02f-99e5-476e-a133-89e959101788 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.666893] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 771.666893] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d7d10f-dd99-0e49-21f3-62b99589aab4" [ 771.666893] env[62235]: _type = "Task" [ 771.666893] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.678855] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d7d10f-dd99-0e49-21f3-62b99589aab4, 'name': SearchDatastore_Task, 'duration_secs': 0.008506} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.679279] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.679601] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] dd7268d1-1da5-4e5f-9d45-816df147b564/dd7268d1-1da5-4e5f-9d45-816df147b564.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 771.680391] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0bed19be-69a2-44c5-9845-7fb712bb4f35 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.688247] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 771.688247] env[62235]: value = "task-1271526" [ 771.688247] env[62235]: _type = "Task" [ 771.688247] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.696680] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.896570] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6858f39a-d554-4bbc-a423-281728e8a805 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.904186] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a77357-9300-4019-8b88-4af0fea39bec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.936943] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b64ed28a-e54d-4d96-a2e3-064f0170dd15 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.946458] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0541b27-3965-463c-94c2-ade9cb74fe28 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.960338] env[62235]: DEBUG nova.compute.provider_tree [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.962443] env[62235]: DEBUG nova.network.neutron [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Updated VIF entry in instance network info cache for port da476e21-90c9-4f68-926c-883707c320ca. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 771.962786] env[62235]: DEBUG nova.network.neutron [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Updating instance_info_cache with network_info: [{"id": "da476e21-90c9-4f68-926c-883707c320ca", "address": "fa:16:3e:22:40:24", "network": {"id": "01e8e9f0-a88d-4080-a81e-8f5597095aef", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-635657508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2171f5abf1ef4e1987b9a77d132215bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52358fcc-0d9f-45dd-8c75-db533fd992c3", "external-id": "nsx-vlan-transportzone-77", "segmentation_id": 77, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda476e21-90", "ovs_interfaceid": "da476e21-90c9-4f68-926c-883707c320ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.149461] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9530d2a7-939c-4a80-859e-99d2d4856269 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "7e6ebba9-58e8-4971-b9fb-8365a61d5505" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.100s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.197838] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427518} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.198124] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] dd7268d1-1da5-4e5f-9d45-816df147b564/dd7268d1-1da5-4e5f-9d45-816df147b564.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 772.198345] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.198588] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f937007-ff77-4483-88b7-2f63aa99a690 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.205560] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 772.205560] env[62235]: value = "task-1271527" [ 772.205560] env[62235]: _type = "Task" [ 772.205560] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.212970] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.465355] env[62235]: DEBUG nova.scheduler.client.report [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.468768] env[62235]: DEBUG oslo_concurrency.lockutils [req-46f3f67c-b8e4-42ed-8fe2-a1f72558b5e5 req-116715df-c3d8-4cc0-a917-f69fafb87aa7 service nova] Releasing lock "refresh_cache-dd7268d1-1da5-4e5f-9d45-816df147b564" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.555782] env[62235]: DEBUG nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.582204] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.582454] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.582608] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.582789] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.582934] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.583091] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.583311] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.583455] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.583618] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.583818] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.583929] env[62235]: DEBUG nova.virt.hardware [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.585423] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2517787-a6a2-4e53-acf1-7c39bca6fe7a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.596719] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9db432-72a5-40bb-87a1-f77248d06a00 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.652507] env[62235]: DEBUG nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.716181] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065522} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.716486] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.717316] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88975e02-3b06-4890-abe8-7efe7e34b930 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.740096] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] dd7268d1-1da5-4e5f-9d45-816df147b564/dd7268d1-1da5-4e5f-9d45-816df147b564.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.740096] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c371974-1366-4b88-9366-888eb7cbdf5f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.761850] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 772.761850] env[62235]: value = "task-1271528" [ 772.761850] env[62235]: _type = "Task" [ 772.761850] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.769786] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271528, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.971289] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.972339] env[62235]: DEBUG nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 772.975061] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.767s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.975297] env[62235]: DEBUG nova.objects.instance [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 773.176351] env[62235]: DEBUG nova.compute.manager [req-e2f26cf6-328e-45aa-8a84-882ea3a16169 req-07c7d7f8-ad72-4a9c-868e-4949dfd5ae5a service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Received event network-vif-plugged-ca4a4847-2fe0-4db9-b4af-43c3870d60a5 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.177160] env[62235]: DEBUG oslo_concurrency.lockutils [req-e2f26cf6-328e-45aa-8a84-882ea3a16169 req-07c7d7f8-ad72-4a9c-868e-4949dfd5ae5a service nova] Acquiring lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.177160] env[62235]: DEBUG oslo_concurrency.lockutils [req-e2f26cf6-328e-45aa-8a84-882ea3a16169 req-07c7d7f8-ad72-4a9c-868e-4949dfd5ae5a service nova] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.177160] env[62235]: DEBUG oslo_concurrency.lockutils [req-e2f26cf6-328e-45aa-8a84-882ea3a16169 req-07c7d7f8-ad72-4a9c-868e-4949dfd5ae5a service nova] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.177512] env[62235]: DEBUG nova.compute.manager [req-e2f26cf6-328e-45aa-8a84-882ea3a16169 req-07c7d7f8-ad72-4a9c-868e-4949dfd5ae5a service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] No waiting events found dispatching network-vif-plugged-ca4a4847-2fe0-4db9-b4af-43c3870d60a5 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 773.177707] env[62235]: WARNING nova.compute.manager [req-e2f26cf6-328e-45aa-8a84-882ea3a16169 req-07c7d7f8-ad72-4a9c-868e-4949dfd5ae5a service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Received unexpected event network-vif-plugged-ca4a4847-2fe0-4db9-b4af-43c3870d60a5 for instance with vm_state building and task_state spawning. [ 773.180171] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.272722] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271528, 'name': ReconfigVM_Task, 'duration_secs': 0.282278} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.273144] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Reconfigured VM instance instance-00000035 to attach disk [datastore2] dd7268d1-1da5-4e5f-9d45-816df147b564/dd7268d1-1da5-4e5f-9d45-816df147b564.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.273838] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc741cf1-6c8b-4879-ac60-c381b40a8b15 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.282108] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 773.282108] env[62235]: value = "task-1271529" [ 773.282108] env[62235]: _type = "Task" [ 773.282108] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.283604] env[62235]: DEBUG nova.network.neutron [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Successfully updated port: ca4a4847-2fe0-4db9-b4af-43c3870d60a5 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 773.296274] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271529, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.482120] env[62235]: DEBUG nova.compute.utils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.482120] env[62235]: DEBUG nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 773.482120] env[62235]: DEBUG nova.network.neutron [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 773.524603] env[62235]: DEBUG nova.policy [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96af7a114d04c0dad3c53f5a2dcd2bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef6547723ccf4be891da50a179dae06f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.789905] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "refresh_cache-d6324b5d-ec70-4bb4-b3c6-b1524abd0034" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.793130] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquired lock "refresh_cache-d6324b5d-ec70-4bb4-b3c6-b1524abd0034" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.793384] env[62235]: DEBUG nova.network.neutron [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.808043] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271529, 'name': Rename_Task, 'duration_secs': 0.132548} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.809428] env[62235]: DEBUG nova.network.neutron [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Successfully created port: 2d93bf16-74a3-4503-8d10-250941a624ff {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.811714] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.812068] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ccf2e66b-67d5-4adb-8db4-92c20b2268aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.819766] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 773.819766] env[62235]: value = "task-1271530" [ 773.819766] env[62235]: _type = "Task" [ 773.819766] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.828600] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271530, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.988917] env[62235]: DEBUG nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.992399] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee56b8de-aca0-498d-a178-6d8a948b88ce tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.993596] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.354s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.995479] env[62235]: INFO nova.compute.claims [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.329746] env[62235]: DEBUG oslo_vmware.api [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271530, 'name': PowerOnVM_Task, 'duration_secs': 0.453989} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.329997] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 774.330205] env[62235]: INFO nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Took 6.11 seconds to spawn the instance on the hypervisor. [ 774.330421] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 774.331201] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af0b8d3-3418-44aa-b2d3-248b63bea03d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.334460] env[62235]: DEBUG nova.network.neutron [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.652465] env[62235]: DEBUG nova.network.neutron [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Updating instance_info_cache with network_info: [{"id": "ca4a4847-2fe0-4db9-b4af-43c3870d60a5", "address": "fa:16:3e:66:1b:eb", "network": {"id": "c9034e31-556b-4292-8464-935c7c006d94", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2119933206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e544511a0a5437faa3d65a8205ed3bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca4a4847-2f", "ovs_interfaceid": "ca4a4847-2fe0-4db9-b4af-43c3870d60a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.851228] env[62235]: INFO nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Took 24.51 seconds to build instance. [ 774.941322] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.941591] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.005099] env[62235]: DEBUG nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.037177] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.037474] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.037939] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.038247] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.038444] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.038638] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.038881] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.039095] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.039307] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.039517] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.039722] env[62235]: DEBUG nova.virt.hardware [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.040720] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf573e2-0795-44bf-a777-e53f76f06bfa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.053019] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd326a48-3e45-4806-8e23-f01dc38e55b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.156725] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Releasing lock "refresh_cache-d6324b5d-ec70-4bb4-b3c6-b1524abd0034" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.156811] env[62235]: DEBUG nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Instance network_info: |[{"id": "ca4a4847-2fe0-4db9-b4af-43c3870d60a5", "address": "fa:16:3e:66:1b:eb", "network": {"id": "c9034e31-556b-4292-8464-935c7c006d94", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2119933206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e544511a0a5437faa3d65a8205ed3bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca4a4847-2f", "ovs_interfaceid": "ca4a4847-2fe0-4db9-b4af-43c3870d60a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 775.157227] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:1b:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31ac3fea-ebf4-4bed-bf70-1eaecdf71280', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca4a4847-2fe0-4db9-b4af-43c3870d60a5', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.164627] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Creating folder: Project (5e544511a0a5437faa3d65a8205ed3bd). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 775.168117] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-532ab947-59e0-4151-8164-40acff9a8d18 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.178584] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Created folder: Project (5e544511a0a5437faa3d65a8205ed3bd) in parent group-v273362. [ 775.179217] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Creating folder: Instances. Parent ref: group-v273393. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 775.179217] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-847f464c-8022-42ba-b89b-13daa3a93809 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.191295] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Created folder: Instances in parent group-v273393. [ 775.191295] env[62235]: DEBUG oslo.service.loopingcall [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.191295] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 775.191295] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e78f9f66-1e74-4944-aa59-cb99884539cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.210636] env[62235]: DEBUG nova.compute.manager [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Received event network-changed-ca4a4847-2fe0-4db9-b4af-43c3870d60a5 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.210829] env[62235]: DEBUG nova.compute.manager [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Refreshing instance network info cache due to event network-changed-ca4a4847-2fe0-4db9-b4af-43c3870d60a5. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 775.211044] env[62235]: DEBUG oslo_concurrency.lockutils [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] Acquiring lock "refresh_cache-d6324b5d-ec70-4bb4-b3c6-b1524abd0034" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.211187] env[62235]: DEBUG oslo_concurrency.lockutils [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] Acquired lock "refresh_cache-d6324b5d-ec70-4bb4-b3c6-b1524abd0034" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.211352] env[62235]: DEBUG nova.network.neutron [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Refreshing network info cache for port ca4a4847-2fe0-4db9-b4af-43c3870d60a5 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 775.213858] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.213858] env[62235]: value = "task-1271533" [ 775.213858] env[62235]: _type = "Task" [ 775.213858] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.222447] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271533, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.352959] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.218s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.355691] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8d53b9-953e-4f19-ac6c-7a5f8171134f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.364268] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01792948-c4ce-4bbe-ba7d-45a770d4150b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.404018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb86b290-8209-4231-b201-25a8f7c8ff49 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.414093] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca8a44b-b3a3-4789-816a-17b4be1043b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.430814] env[62235]: DEBUG nova.compute.provider_tree [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.474282] env[62235]: DEBUG nova.network.neutron [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Successfully updated port: 2d93bf16-74a3-4503-8d10-250941a624ff {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 775.726907] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271533, 'name': CreateVM_Task, 'duration_secs': 0.310661} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.727246] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 775.727781] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.727942] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.728281] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 775.728545] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7708df32-1e89-4978-be02-104a84b81273 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.733275] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 775.733275] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e042b4-ae47-5361-6be4-7450cc708453" [ 775.733275] env[62235]: _type = "Task" [ 775.733275] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.742121] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e042b4-ae47-5361-6be4-7450cc708453, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.861586] env[62235]: DEBUG nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 775.933567] env[62235]: DEBUG nova.scheduler.client.report [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.954331] env[62235]: DEBUG nova.network.neutron [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Updated VIF entry in instance network info cache for port ca4a4847-2fe0-4db9-b4af-43c3870d60a5. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 775.954683] env[62235]: DEBUG nova.network.neutron [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Updating instance_info_cache with network_info: [{"id": "ca4a4847-2fe0-4db9-b4af-43c3870d60a5", "address": "fa:16:3e:66:1b:eb", "network": {"id": "c9034e31-556b-4292-8464-935c7c006d94", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2119933206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e544511a0a5437faa3d65a8205ed3bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca4a4847-2f", "ovs_interfaceid": "ca4a4847-2fe0-4db9-b4af-43c3870d60a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.975131] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-afa35ae0-40c6-4eaa-80f5-c1ada27f1871" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.975293] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-afa35ae0-40c6-4eaa-80f5-c1ada27f1871" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.975570] env[62235]: DEBUG nova.network.neutron [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.244522] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e042b4-ae47-5361-6be4-7450cc708453, 'name': SearchDatastore_Task, 'duration_secs': 0.011585} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.245121] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.245360] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.245592] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.245735] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.245911] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.246192] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29a3370f-2008-4d5a-ac8b-35b6da863bd3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.253918] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.254107] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 776.254795] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f0bf9f-e396-4cee-b60b-f393a6f49617 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.259704] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 776.259704] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52731dfa-cb84-03ab-3cd1-c35aecb9974c" [ 776.259704] env[62235]: _type = "Task" [ 776.259704] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.267786] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52731dfa-cb84-03ab-3cd1-c35aecb9974c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.381186] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.440325] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.440864] env[62235]: DEBUG nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.443475] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.565s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.457202] env[62235]: DEBUG oslo_concurrency.lockutils [req-c8805791-96e4-4ffc-9eca-fdd1360798f5 req-5d13e824-69a7-446a-9b96-eebb8aebd22d service nova] Releasing lock "refresh_cache-d6324b5d-ec70-4bb4-b3c6-b1524abd0034" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.504410] env[62235]: DEBUG nova.network.neutron [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.663054] env[62235]: DEBUG nova.network.neutron [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Updating instance_info_cache with network_info: [{"id": "2d93bf16-74a3-4503-8d10-250941a624ff", "address": "fa:16:3e:46:17:0b", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d93bf16-74", "ovs_interfaceid": "2d93bf16-74a3-4503-8d10-250941a624ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.771680] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52731dfa-cb84-03ab-3cd1-c35aecb9974c, 'name': SearchDatastore_Task, 'duration_secs': 0.007907} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.772475] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e63d692a-caeb-4732-9684-21cd4daa1d2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.777891] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 776.777891] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b651e1-8767-5cd5-88e2-923db00e5850" [ 776.777891] env[62235]: _type = "Task" [ 776.777891] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.785488] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b651e1-8767-5cd5-88e2-923db00e5850, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.953116] env[62235]: DEBUG nova.compute.utils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.956682] env[62235]: DEBUG nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Not allocating networking since 'none' was specified. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 777.168248] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-afa35ae0-40c6-4eaa-80f5-c1ada27f1871" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.168544] env[62235]: DEBUG nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Instance network_info: |[{"id": "2d93bf16-74a3-4503-8d10-250941a624ff", "address": "fa:16:3e:46:17:0b", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d93bf16-74", "ovs_interfaceid": "2d93bf16-74a3-4503-8d10-250941a624ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 777.168933] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:17:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d93bf16-74a3-4503-8d10-250941a624ff', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.177172] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating folder: Project (ef6547723ccf4be891da50a179dae06f). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.178861] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ebbbf48-b038-4ad5-98ec-4af82919b467 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.196045] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created folder: Project (ef6547723ccf4be891da50a179dae06f) in parent group-v273362. [ 777.196045] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating folder: Instances. Parent ref: group-v273396. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.196045] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7fa76c54-a549-4056-b3c5-0d49447453d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.205699] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created folder: Instances in parent group-v273396. [ 777.205999] env[62235]: DEBUG oslo.service.loopingcall [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.208384] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.208774] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22d72f29-a62c-4618-ab55-823a27a25143 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.230789] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.230789] env[62235]: value = "task-1271536" [ 777.230789] env[62235]: _type = "Task" [ 777.230789] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.238682] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271536, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.244860] env[62235]: DEBUG nova.compute.manager [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Received event network-vif-plugged-2d93bf16-74a3-4503-8d10-250941a624ff {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.244860] env[62235]: DEBUG oslo_concurrency.lockutils [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] Acquiring lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.244860] env[62235]: DEBUG oslo_concurrency.lockutils [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.244860] env[62235]: DEBUG oslo_concurrency.lockutils [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.245025] env[62235]: DEBUG nova.compute.manager [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] No waiting events found dispatching network-vif-plugged-2d93bf16-74a3-4503-8d10-250941a624ff {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 777.245178] env[62235]: WARNING nova.compute.manager [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Received unexpected event network-vif-plugged-2d93bf16-74a3-4503-8d10-250941a624ff for instance with vm_state building and task_state spawning. [ 777.245371] env[62235]: DEBUG nova.compute.manager [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Received event network-changed-2d93bf16-74a3-4503-8d10-250941a624ff {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.245536] env[62235]: DEBUG nova.compute.manager [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Refreshing instance network info cache due to event network-changed-2d93bf16-74a3-4503-8d10-250941a624ff. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 777.245717] env[62235]: DEBUG oslo_concurrency.lockutils [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] Acquiring lock "refresh_cache-afa35ae0-40c6-4eaa-80f5-c1ada27f1871" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.245850] env[62235]: DEBUG oslo_concurrency.lockutils [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] Acquired lock "refresh_cache-afa35ae0-40c6-4eaa-80f5-c1ada27f1871" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.246010] env[62235]: DEBUG nova.network.neutron [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Refreshing network info cache for port 2d93bf16-74a3-4503-8d10-250941a624ff {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 777.267039] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f282fac7-b077-4e8e-b1e6-6d8969a13265 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.274542] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ee69f6-600f-49e6-8b44-aa5945a4c0a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.308628] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2705a412-0aa6-429c-a6f6-302d84e54230 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.315240] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b651e1-8767-5cd5-88e2-923db00e5850, 'name': SearchDatastore_Task, 'duration_secs': 0.008939} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.315618] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.315911] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d6324b5d-ec70-4bb4-b3c6-b1524abd0034/d6324b5d-ec70-4bb4-b3c6-b1524abd0034.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 777.316743] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf42f796-f12b-4246-8d60-0ac43c28cfbb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.324036] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cdb8fd9-1102-4d77-b9b6-8ab7a2f35ec4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.327914] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 777.327914] env[62235]: value = "task-1271537" [ 777.327914] env[62235]: _type = "Task" [ 777.327914] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.339082] env[62235]: DEBUG nova.compute.provider_tree [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.344864] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271537, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.456886] env[62235]: DEBUG nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.741394] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271536, 'name': CreateVM_Task, 'duration_secs': 0.460431} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.741674] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 777.742432] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.742602] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.742879] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.743213] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ff57085-f9b3-4f59-827e-2d1aab0230dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.750611] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 777.750611] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528bf36f-af8f-3e59-1e2c-c06f03f8f246" [ 777.750611] env[62235]: _type = "Task" [ 777.750611] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.759673] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528bf36f-af8f-3e59-1e2c-c06f03f8f246, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.837130] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271537, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454672} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.839291] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d6324b5d-ec70-4bb4-b3c6-b1524abd0034/d6324b5d-ec70-4bb4-b3c6-b1524abd0034.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 777.839512] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 777.839767] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f380a66-24a0-47fa-84d4-72a25c4651be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.842348] env[62235]: DEBUG nova.scheduler.client.report [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.850343] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 777.850343] env[62235]: value = "task-1271538" [ 777.850343] env[62235]: _type = "Task" [ 777.850343] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.857735] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.957441] env[62235]: DEBUG nova.network.neutron [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Updated VIF entry in instance network info cache for port 2d93bf16-74a3-4503-8d10-250941a624ff. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 777.957809] env[62235]: DEBUG nova.network.neutron [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Updating instance_info_cache with network_info: [{"id": "2d93bf16-74a3-4503-8d10-250941a624ff", "address": "fa:16:3e:46:17:0b", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d93bf16-74", "ovs_interfaceid": "2d93bf16-74a3-4503-8d10-250941a624ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.261277] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528bf36f-af8f-3e59-1e2c-c06f03f8f246, 'name': SearchDatastore_Task, 'duration_secs': 0.02128} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.261573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.261805] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.262044] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.262210] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.262397] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.262652] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3587de36-c707-48cf-84fd-135e12297adb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.271917] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.272111] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 778.272808] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b9b0f1e-4195-4c41-a3a0-730327ebc614 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.278135] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 778.278135] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5292582e-7ab7-42c1-8c76-262e721bbf07" [ 778.278135] env[62235]: _type = "Task" [ 778.278135] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.285152] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5292582e-7ab7-42c1-8c76-262e721bbf07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.346889] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.347532] env[62235]: ERROR nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Traceback (most recent call last): [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self.driver.spawn(context, instance, image_meta, [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] vm_ref = self.build_virtual_machine(instance, [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.347532] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] for vif in network_info: [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return self._sync_wrapper(fn, *args, **kwargs) [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self.wait() [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self[:] = self._gt.wait() [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return self._exit_event.wait() [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] result = hub.switch() [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 778.348047] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return self.greenlet.switch() [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] result = function(*args, **kwargs) [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] return func(*args, **kwargs) [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] raise e [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] nwinfo = self.network_api.allocate_for_instance( [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] created_port_ids = self._update_ports_for_instance( [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] with excutils.save_and_reraise_exception(): [ 778.348706] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] self.force_reraise() [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] raise self.value [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] updated_port = self._update_port( [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] _ensure_no_port_binding_failure(port) [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] raise exception.PortBindingFailed(port_id=port['id']) [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] nova.exception.PortBindingFailed: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. [ 778.349277] env[62235]: ERROR nova.compute.manager [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] [ 778.349649] env[62235]: DEBUG nova.compute.utils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 778.349649] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.432s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.352460] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Build of instance 6671163c-c829-4046-97e5-e85fe5e223dd was re-scheduled: Binding failed for port 47d535bf-048e-404c-825f-0290d9fce21f, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 778.352849] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 778.353089] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Acquiring lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.353240] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Acquired lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.353399] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.362877] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063841} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.363608] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.364456] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542f7c0d-02c3-4c7d-94a4-4d75621c7d5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.389208] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] d6324b5d-ec70-4bb4-b3c6-b1524abd0034/d6324b5d-ec70-4bb4-b3c6-b1524abd0034.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.390020] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5903e2f1-ebb3-45aa-8ec2-6bca87616e49 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.411343] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 778.411343] env[62235]: value = "task-1271539" [ 778.411343] env[62235]: _type = "Task" [ 778.411343] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.419231] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.460455] env[62235]: DEBUG oslo_concurrency.lockutils [req-7db0df6f-fd9e-45d4-b870-dd4b39055354 req-bf6b0491-8f90-4601-ba47-11e68c7b53a4 service nova] Releasing lock "refresh_cache-afa35ae0-40c6-4eaa-80f5-c1ada27f1871" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.466838] env[62235]: DEBUG nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.493686] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.493957] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.494132] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.494383] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.494538] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.494689] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.494915] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.495059] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.495237] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.495429] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.495595] env[62235]: DEBUG nova.virt.hardware [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.496499] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985801ea-852e-4cbe-84eb-9bf64b98d5b4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.506024] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bcbb3a-b5c2-4858-8180-441585745ee6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.517445] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.522760] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Creating folder: Project (88f449ab37074e7080aebcd6df090863). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.523051] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e78567e2-5f40-4e54-a651-c5ec7f1e47a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.532557] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Created folder: Project (88f449ab37074e7080aebcd6df090863) in parent group-v273362. [ 778.532739] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Creating folder: Instances. Parent ref: group-v273399. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.532955] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f1bd039-472e-408e-8dae-29f77b9b9ece {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.541494] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Created folder: Instances in parent group-v273399. [ 778.541719] env[62235]: DEBUG oslo.service.loopingcall [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.541896] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.542103] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd3fd7b9-5860-46a4-8e0f-9ba533b26f76 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.557440] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.557440] env[62235]: value = "task-1271542" [ 778.557440] env[62235]: _type = "Task" [ 778.557440] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.564350] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271542, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.788232] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5292582e-7ab7-42c1-8c76-262e721bbf07, 'name': SearchDatastore_Task, 'duration_secs': 0.007999} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.788864] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6582bb9d-ec6b-41cb-8376-e437ed47a2eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.793842] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 778.793842] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a9fa59-80ef-5a6b-2430-de5410c687d3" [ 778.793842] env[62235]: _type = "Task" [ 778.793842] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.801059] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a9fa59-80ef-5a6b-2430-de5410c687d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.876520] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.922377] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271539, 'name': ReconfigVM_Task, 'duration_secs': 0.278313} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.924487] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Reconfigured VM instance instance-00000036 to attach disk [datastore1] d6324b5d-ec70-4bb4-b3c6-b1524abd0034/d6324b5d-ec70-4bb4-b3c6-b1524abd0034.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.925302] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7500d3f4-7f38-4012-a82e-f8f2c5b2eaa6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.931473] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 778.931473] env[62235]: value = "task-1271543" [ 778.931473] env[62235]: _type = "Task" [ 778.931473] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.939132] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271543, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.969019] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.066091] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271542, 'name': CreateVM_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.135742] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10212ff6-5e9e-4077-a5eb-4381ac98f838 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.144702] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3d9c8c-d9a5-4b14-bd90-68d444b4ac41 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.189380] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b043fe5d-02d2-4ed7-bc40-02d3db5f8244 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.197480] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be017c25-5bb1-4fea-8257-266d760b7960 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.212547] env[62235]: DEBUG nova.compute.provider_tree [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.304426] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a9fa59-80ef-5a6b-2430-de5410c687d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008517} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.304711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.304972] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] afa35ae0-40c6-4eaa-80f5-c1ada27f1871/afa35ae0-40c6-4eaa-80f5-c1ada27f1871.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.305252] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12444a4c-dfaf-40c9-acea-07204b2ebf09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.312342] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 779.312342] env[62235]: value = "task-1271544" [ 779.312342] env[62235]: _type = "Task" [ 779.312342] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.321217] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.444123] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271543, 'name': Rename_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.470959] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Releasing lock "refresh_cache-6671163c-c829-4046-97e5-e85fe5e223dd" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.471287] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 779.471479] env[62235]: DEBUG nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.471654] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 779.487969] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.570625] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271542, 'name': CreateVM_Task, 'duration_secs': 0.789764} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.570778] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.571424] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.571650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.572024] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.572717] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e452b67d-7160-41ca-954b-43e7b593e583 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.577820] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 779.577820] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52522e53-6429-a965-0584-638c8fb4d821" [ 779.577820] env[62235]: _type = "Task" [ 779.577820] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.587366] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52522e53-6429-a965-0584-638c8fb4d821, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.715575] env[62235]: DEBUG nova.scheduler.client.report [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.824516] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.422638} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.824859] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] afa35ae0-40c6-4eaa-80f5-c1ada27f1871/afa35ae0-40c6-4eaa-80f5-c1ada27f1871.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 779.825099] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 779.825364] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8791b250-3122-496d-a968-cc32ca38afbe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.832569] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 779.832569] env[62235]: value = "task-1271545" [ 779.832569] env[62235]: _type = "Task" [ 779.832569] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.840112] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271545, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.942893] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271543, 'name': Rename_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.990613] env[62235]: DEBUG nova.network.neutron [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.087191] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52522e53-6429-a965-0584-638c8fb4d821, 'name': SearchDatastore_Task, 'duration_secs': 0.053405} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.087492] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.087720] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.087950] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.088107] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.088285] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.088558] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe541180-b403-4243-80ad-d392421c4059 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.096138] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.096321] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 780.097014] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d46fad46-e5b8-4af8-8ecf-ce1a329ed7ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.102616] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 780.102616] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52949edf-e1a4-4fc8-6465-03f53d9d9df8" [ 780.102616] env[62235]: _type = "Task" [ 780.102616] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.110137] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52949edf-e1a4-4fc8-6465-03f53d9d9df8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.220928] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.871s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.221651] env[62235]: ERROR nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Traceback (most recent call last): [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self.driver.spawn(context, instance, image_meta, [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] vm_ref = self.build_virtual_machine(instance, [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.221651] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] for vif in network_info: [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return self._sync_wrapper(fn, *args, **kwargs) [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self.wait() [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self[:] = self._gt.wait() [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return self._exit_event.wait() [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] result = hub.switch() [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 780.222025] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return self.greenlet.switch() [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] result = function(*args, **kwargs) [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] return func(*args, **kwargs) [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] raise e [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] nwinfo = self.network_api.allocate_for_instance( [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] created_port_ids = self._update_ports_for_instance( [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] with excutils.save_and_reraise_exception(): [ 780.222377] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] self.force_reraise() [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] raise self.value [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] updated_port = self._update_port( [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] _ensure_no_port_binding_failure(port) [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] raise exception.PortBindingFailed(port_id=port['id']) [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] nova.exception.PortBindingFailed: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. [ 780.222763] env[62235]: ERROR nova.compute.manager [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] [ 780.223056] env[62235]: DEBUG nova.compute.utils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.223825] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.069s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.224099] env[62235]: DEBUG nova.objects.instance [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lazy-loading 'resources' on Instance uuid 430056ca-ccba-45c5-adf7-9c068f81f9ec {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 780.225964] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Build of instance 3e7a81ae-2163-4562-b483-dc639c602fa0 was re-scheduled: Binding failed for port 9cad198b-7619-4450-8e94-47b42ba1d8a6, please check neutron logs for more information. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 780.225964] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 780.226150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.226319] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquired lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.226517] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.343033] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271545, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059351} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.343033] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.343275] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d194f40-2739-436d-b8a3-c3f37e0c60e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.364891] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] afa35ae0-40c6-4eaa-80f5-c1ada27f1871/afa35ae0-40c6-4eaa-80f5-c1ada27f1871.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.365556] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db8ef261-ac94-4c2f-8c07-fab984bc05e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.384861] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 780.384861] env[62235]: value = "task-1271546" [ 780.384861] env[62235]: _type = "Task" [ 780.384861] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.392655] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.444121] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271543, 'name': Rename_Task, 'duration_secs': 1.143335} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.444121] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 780.444121] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f66ee2b4-2f41-471f-a9ca-85e26a1d212e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.450032] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 780.450032] env[62235]: value = "task-1271547" [ 780.450032] env[62235]: _type = "Task" [ 780.450032] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.457823] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271547, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.493777] env[62235]: INFO nova.compute.manager [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] [instance: 6671163c-c829-4046-97e5-e85fe5e223dd] Took 1.02 seconds to deallocate network for instance. [ 780.613437] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52949edf-e1a4-4fc8-6465-03f53d9d9df8, 'name': SearchDatastore_Task, 'duration_secs': 0.007647} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.614250] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a2cc7e1-5a86-4c41-bbb1-55f240eef403 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.619236] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 780.619236] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e035a-6429-0fe0-bf17-08b7219edba1" [ 780.619236] env[62235]: _type = "Task" [ 780.619236] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.628543] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e035a-6429-0fe0-bf17-08b7219edba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.745906] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.748434] env[62235]: DEBUG nova.scheduler.client.report [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 780.763678] env[62235]: DEBUG nova.scheduler.client.report [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 780.763921] env[62235]: DEBUG nova.compute.provider_tree [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 780.775184] env[62235]: DEBUG nova.scheduler.client.report [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 781.536434] env[62235]: DEBUG nova.scheduler.client.report [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 781.543611] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.558726] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271546, 'name': ReconfigVM_Task, 'duration_secs': 0.253962} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.564055] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Reconfigured VM instance instance-00000037 to attach disk [datastore1] afa35ae0-40c6-4eaa-80f5-c1ada27f1871/afa35ae0-40c6-4eaa-80f5-c1ada27f1871.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.564966] env[62235]: DEBUG oslo_vmware.api [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271547, 'name': PowerOnVM_Task, 'duration_secs': 0.43225} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.565165] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e035a-6429-0fe0-bf17-08b7219edba1, 'name': SearchDatastore_Task, 'duration_secs': 0.009066} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.565362] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef889bd3-e29f-4dd0-a347-c4b6251d7140 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.566876] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 781.567088] env[62235]: INFO nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Took 9.01 seconds to spawn the instance on the hypervisor. [ 781.567597] env[62235]: DEBUG nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 781.567597] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.567740] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] ec1830c7-7095-45d1-a188-c2993c60035e/ec1830c7-7095-45d1-a188-c2993c60035e.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 781.568535] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd1a89a-c603-46e0-88ca-e87fa4e5e437 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.570700] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52becf44-77a9-46cc-8b7b-3d3682409899 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.581335] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 781.581335] env[62235]: value = "task-1271549" [ 781.581335] env[62235]: _type = "Task" [ 781.581335] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.581576] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 781.581576] env[62235]: value = "task-1271548" [ 781.581576] env[62235]: _type = "Task" [ 781.581576] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.592610] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271549, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.595677] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271548, 'name': Rename_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.038358] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af7925d-2974-4a7e-b380-22f5528c7118 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.049117] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66aef89f-da48-425b-b1c6-fe2e6269a755 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.053579] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Releasing lock "refresh_cache-3e7a81ae-2163-4562-b483-dc639c602fa0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.053806] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.053985] env[62235]: DEBUG nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.054165] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.089396] env[62235]: INFO nova.scheduler.client.report [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Deleted allocations for instance 6671163c-c829-4046-97e5-e85fe5e223dd [ 782.101198] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789491ba-4ecd-4875-9a8b-5ebd73aae80d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.106187] env[62235]: INFO nova.compute.manager [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Took 26.24 seconds to build instance. [ 782.108500] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.116544] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271549, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.116854] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271548, 'name': Rename_Task, 'duration_secs': 0.159637} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.118631] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbb75da-4b68-4b49-aa90-21ef848de9cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.122721] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.123415] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad71364f-9757-4699-a12a-cfa5e4db0cca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.135411] env[62235]: DEBUG nova.compute.provider_tree [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.137388] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 782.137388] env[62235]: value = "task-1271550" [ 782.137388] env[62235]: _type = "Task" [ 782.137388] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.146502] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.599140] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271549, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544738} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.599559] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] ec1830c7-7095-45d1-a188-c2993c60035e/ec1830c7-7095-45d1-a188-c2993c60035e.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 782.599700] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.600036] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8f846c5-70d0-404d-abfd-d8b03b542571 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.607991] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 782.607991] env[62235]: value = "task-1271551" [ 782.607991] env[62235]: _type = "Task" [ 782.607991] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.608339] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e5e8699-573a-4019-8b3b-f14ed69031c2 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.339s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.608780] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53a67b85-2a76-4595-864c-ca5c52c97e66 tempest-ServersTestFqdnHostnames-1305387895 tempest-ServersTestFqdnHostnames-1305387895-project-member] Lock "6671163c-c829-4046-97e5-e85fe5e223dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.373s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.617635] env[62235]: DEBUG nova.network.neutron [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.623477] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271551, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.641028] env[62235]: DEBUG nova.scheduler.client.report [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.652200] env[62235]: DEBUG oslo_vmware.api [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271550, 'name': PowerOnVM_Task, 'duration_secs': 0.510555} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.652321] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 782.652648] env[62235]: INFO nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Took 7.65 seconds to spawn the instance on the hypervisor. [ 782.652963] env[62235]: DEBUG nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.653863] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf39802-fad5-4f9f-bf63-4ac5fb1ec55a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.118036] env[62235]: DEBUG nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.121202] env[62235]: DEBUG nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.124303] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271551, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092851} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.124303] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.124303] env[62235]: INFO nova.compute.manager [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: 3e7a81ae-2163-4562-b483-dc639c602fa0] Took 1.07 seconds to deallocate network for instance. [ 783.128000] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602cba96-d691-49c7-a9dc-15adae356f37 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.158436] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] ec1830c7-7095-45d1-a188-c2993c60035e/ec1830c7-7095-45d1-a188-c2993c60035e.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.162027] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.934s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.162350] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03f85959-8e2f-406f-a37c-52dcd2dff69c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.193378] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.652s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.198566] env[62235]: INFO nova.compute.claims [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.212455] env[62235]: INFO nova.compute.manager [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Took 25.52 seconds to build instance. [ 783.215205] env[62235]: INFO nova.scheduler.client.report [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Deleted allocations for instance 430056ca-ccba-45c5-adf7-9c068f81f9ec [ 783.216363] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 783.216363] env[62235]: value = "task-1271552" [ 783.216363] env[62235]: _type = "Task" [ 783.216363] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.231822] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271552, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.649433] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.650357] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.721026] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8032a4d6-4be4-4f45-80a2-556b810bf7ab tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.466s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.730917] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d5016683-51ce-4069-a006-e6268b1e1271 tempest-ServerShowV254Test-8312804 tempest-ServerShowV254Test-8312804-project-member] Lock "430056ca-ccba-45c5-adf7-9c068f81f9ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.369s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.736681] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271552, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.041403] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.042041] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.042647] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.042647] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.042769] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.045190] env[62235]: INFO nova.compute.manager [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Terminating instance [ 784.047034] env[62235]: DEBUG nova.compute.manager [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.047233] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.048928] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7230813-cc27-4906-90ab-1394e321efd6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.056497] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 784.056732] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3eadb21a-7ca3-4c6c-86dc-2065ea35ddc9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.062851] env[62235]: DEBUG oslo_vmware.api [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 784.062851] env[62235]: value = "task-1271553" [ 784.062851] env[62235]: _type = "Task" [ 784.062851] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.070621] env[62235]: DEBUG oslo_vmware.api [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.160837] env[62235]: INFO nova.scheduler.client.report [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Deleted allocations for instance 3e7a81ae-2163-4562-b483-dc639c602fa0 [ 784.227603] env[62235]: DEBUG nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.236127] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271552, 'name': ReconfigVM_Task, 'duration_secs': 0.606405} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.239536] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Reconfigured VM instance instance-00000038 to attach disk [datastore1] ec1830c7-7095-45d1-a188-c2993c60035e/ec1830c7-7095-45d1-a188-c2993c60035e.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.240457] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19d3aa0f-2f26-4e82-9d3b-6b7507894782 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.249025] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 784.249025] env[62235]: value = "task-1271554" [ 784.249025] env[62235]: _type = "Task" [ 784.249025] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.260858] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271554, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.473941] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.473941] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.473941] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.473941] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.474465] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.477727] env[62235]: INFO nova.compute.manager [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Terminating instance [ 784.479789] env[62235]: DEBUG nova.compute.manager [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.480106] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.483018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867325c2-f3df-497a-a9b8-fa1f7eb35f69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.493122] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 784.493122] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de34eceb-69a2-4381-8637-912fa9d83c8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.504019] env[62235]: DEBUG oslo_vmware.api [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 784.504019] env[62235]: value = "task-1271555" [ 784.504019] env[62235]: _type = "Task" [ 784.504019] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.509676] env[62235]: DEBUG oslo_vmware.api [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.526048] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212f61ec-2037-4f32-bb9e-4ae57e97953c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.533216] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e38a5e2-1c79-4241-841a-f98bd4dc22e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.571545] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcaa1642-abe3-4306-adaf-467b510a1569 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.579189] env[62235]: DEBUG oslo_vmware.api [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271553, 'name': PowerOffVM_Task, 'duration_secs': 0.212609} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.584746] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 784.584746] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 784.584746] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29eb3869-3a2a-41a5-b0a5-f492b499eab8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.585516] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749478e5-b2dd-43ca-ba7d-af37317d5ba4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.601416] env[62235]: DEBUG nova.compute.provider_tree [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.649559] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 784.649820] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 784.649937] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Deleting the datastore file [datastore1] d6324b5d-ec70-4bb4-b3c6-b1524abd0034 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.650219] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18eff3c6-722a-479e-a03b-3ed0d791a95f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.657502] env[62235]: DEBUG oslo_vmware.api [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for the task: (returnval){ [ 784.657502] env[62235]: value = "task-1271557" [ 784.657502] env[62235]: _type = "Task" [ 784.657502] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.666228] env[62235]: DEBUG oslo_vmware.api [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271557, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.669777] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a41a4f3a-8f21-45cc-a49a-4400a005f509 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "3e7a81ae-2163-4562-b483-dc639c602fa0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.565s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.757481] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.761247] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271554, 'name': Rename_Task, 'duration_secs': 0.133857} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.761646] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 784.761923] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-302b69a5-9f55-419c-9200-ddbc6357a158 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.768550] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 784.768550] env[62235]: value = "task-1271558" [ 784.768550] env[62235]: _type = "Task" [ 784.768550] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.777225] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271558, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.017620] env[62235]: DEBUG oslo_vmware.api [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271555, 'name': PowerOffVM_Task, 'duration_secs': 0.305468} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.018674] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 785.018958] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 785.019280] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d64de0f2-111f-4c53-8b3a-27b2428c1199 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.083318] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 785.083318] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 785.083318] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleting the datastore file [datastore1] afa35ae0-40c6-4eaa-80f5-c1ada27f1871 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.083318] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eac95dc8-fad7-44af-a737-1e71486f7c34 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.089672] env[62235]: DEBUG oslo_vmware.api [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 785.089672] env[62235]: value = "task-1271560" [ 785.089672] env[62235]: _type = "Task" [ 785.089672] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.097170] env[62235]: DEBUG oslo_vmware.api [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271560, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.104247] env[62235]: DEBUG nova.scheduler.client.report [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.168864] env[62235]: DEBUG oslo_vmware.api [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Task: {'id': task-1271557, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22928} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.169201] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.169388] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 785.169560] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.169855] env[62235]: INFO nova.compute.manager [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Took 1.12 seconds to destroy the instance on the hypervisor. [ 785.169961] env[62235]: DEBUG oslo.service.loopingcall [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.170168] env[62235]: DEBUG nova.compute.manager [-] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.170257] env[62235]: DEBUG nova.network.neutron [-] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 785.172478] env[62235]: DEBUG nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.278580] env[62235]: DEBUG oslo_vmware.api [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271558, 'name': PowerOnVM_Task, 'duration_secs': 0.427537} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.278849] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 785.279054] env[62235]: INFO nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Took 6.81 seconds to spawn the instance on the hypervisor. [ 785.279231] env[62235]: DEBUG nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.279989] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6201ff-8433-4413-9842-d2e7c40b3ffb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.412344] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "dd7268d1-1da5-4e5f-9d45-816df147b564" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.412667] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.413071] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "dd7268d1-1da5-4e5f-9d45-816df147b564-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.413071] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.413246] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.415476] env[62235]: INFO nova.compute.manager [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Terminating instance [ 785.417758] env[62235]: DEBUG nova.compute.manager [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 785.417958] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.418876] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269ee471-3676-4779-991f-facf61e7851b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.426532] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 785.426778] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f4bb05d-fe99-44ce-82d7-c18f6d03292b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.434090] env[62235]: DEBUG oslo_vmware.api [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 785.434090] env[62235]: value = "task-1271561" [ 785.434090] env[62235]: _type = "Task" [ 785.434090] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.444230] env[62235]: DEBUG oslo_vmware.api [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271561, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.556725] env[62235]: DEBUG nova.compute.manager [req-0081a3e2-aed4-4f59-b55e-f0aa5e36df03 req-7e33b79d-7c39-4f57-b275-b24286de41cf service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Received event network-vif-deleted-ca4a4847-2fe0-4db9-b4af-43c3870d60a5 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.556948] env[62235]: INFO nova.compute.manager [req-0081a3e2-aed4-4f59-b55e-f0aa5e36df03 req-7e33b79d-7c39-4f57-b275-b24286de41cf service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Neutron deleted interface ca4a4847-2fe0-4db9-b4af-43c3870d60a5; detaching it from the instance and deleting it from the info cache [ 785.557161] env[62235]: DEBUG nova.network.neutron [req-0081a3e2-aed4-4f59-b55e-f0aa5e36df03 req-7e33b79d-7c39-4f57-b275-b24286de41cf service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.599088] env[62235]: DEBUG oslo_vmware.api [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271560, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199913} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.599370] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.599601] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 785.599722] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.599894] env[62235]: INFO nova.compute.manager [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Took 1.12 seconds to destroy the instance on the hypervisor. [ 785.600144] env[62235]: DEBUG oslo.service.loopingcall [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.600336] env[62235]: DEBUG nova.compute.manager [-] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.600430] env[62235]: DEBUG nova.network.neutron [-] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 785.609573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.610180] env[62235]: DEBUG nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.613884] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.346s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.695721] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.803502] env[62235]: INFO nova.compute.manager [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Took 27.18 seconds to build instance. [ 785.951442] env[62235]: DEBUG oslo_vmware.api [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271561, 'name': PowerOffVM_Task, 'duration_secs': 0.489672} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.952710] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 785.952930] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 785.953205] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7416ec49-3ff5-4356-b377-a7907ff256c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.987704] env[62235]: DEBUG nova.network.neutron [-] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.024561] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 786.024760] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 786.025437] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Deleting the datastore file [datastore2] dd7268d1-1da5-4e5f-9d45-816df147b564 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.027996] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab68285d-f9ed-45f7-82b5-df03412f4858 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.033631] env[62235]: DEBUG oslo_vmware.api [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for the task: (returnval){ [ 786.033631] env[62235]: value = "task-1271563" [ 786.033631] env[62235]: _type = "Task" [ 786.033631] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.043867] env[62235]: DEBUG oslo_vmware.api [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271563, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.060351] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a81cde32-0890-4c5c-badb-75d7f502c656 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.071422] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e093be-091d-4391-84e7-8dfa56ddc8b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.096847] env[62235]: DEBUG nova.compute.manager [req-0081a3e2-aed4-4f59-b55e-f0aa5e36df03 req-7e33b79d-7c39-4f57-b275-b24286de41cf service nova] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Detach interface failed, port_id=ca4a4847-2fe0-4db9-b4af-43c3870d60a5, reason: Instance d6324b5d-ec70-4bb4-b3c6-b1524abd0034 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 786.121139] env[62235]: DEBUG nova.compute.utils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.122433] env[62235]: DEBUG nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Not allocating networking since 'none' was specified. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 786.305054] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f4ccf46-f1c2-4761-89e6-309a0ec4332e tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "ec1830c7-7095-45d1-a188-c2993c60035e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.797s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.491330] env[62235]: INFO nova.compute.manager [-] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Took 1.32 seconds to deallocate network for instance. [ 786.531021] env[62235]: DEBUG nova.network.neutron [-] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.542875] env[62235]: DEBUG oslo_vmware.api [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Task: {'id': task-1271563, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136712} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.543460] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.543559] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 786.543743] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.544714] env[62235]: INFO nova.compute.manager [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Took 1.13 seconds to destroy the instance on the hypervisor. [ 786.544714] env[62235]: DEBUG oslo.service.loopingcall [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.544714] env[62235]: DEBUG nova.compute.manager [-] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.544714] env[62235]: DEBUG nova.network.neutron [-] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.624408] env[62235]: DEBUG nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.647466] env[62235]: WARNING nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 786.648206] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance dd7268d1-1da5-4e5f-9d45-816df147b564 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.648206] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance d6324b5d-ec70-4bb4-b3c6-b1524abd0034 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.648206] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance afa35ae0-40c6-4eaa-80f5-c1ada27f1871 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.648206] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ec1830c7-7095-45d1-a188-c2993c60035e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.648354] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b4a3777f-2ee5-464b-84eb-0722d59f2b44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.808343] env[62235]: DEBUG nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.997947] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.035818] env[62235]: INFO nova.compute.manager [-] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Took 1.44 seconds to deallocate network for instance. [ 787.151650] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3ea7db36-ebde-4523-b803-ad121a00269b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.288184] env[62235]: DEBUG nova.network.neutron [-] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.330917] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.543275] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.641752] env[62235]: DEBUG nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.657690] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 63e16fd4-3a5a-4231-9013-c6a01f0eab80 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.668485] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.668761] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.668950] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.669193] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.669367] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.669526] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.669781] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.669974] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.670310] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.670516] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.670723] env[62235]: DEBUG nova.virt.hardware [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.671674] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f044cb31-73ef-449f-9bbe-96fd2de64be3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.681118] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89a016a-eb32-438f-8afb-0e886ce5c08f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.696126] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.701719] env[62235]: DEBUG oslo.service.loopingcall [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.701976] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.702203] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64de4447-a4f7-40de-9177-3ea20e26532d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.716235] env[62235]: DEBUG nova.compute.manager [req-e56ae770-b041-4ed7-b49f-c7f4959da654 req-e48e2956-0af0-43ac-b692-f90ee129794d service nova] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Received event network-vif-deleted-2d93bf16-74a3-4503-8d10-250941a624ff {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.716449] env[62235]: DEBUG nova.compute.manager [req-e56ae770-b041-4ed7-b49f-c7f4959da654 req-e48e2956-0af0-43ac-b692-f90ee129794d service nova] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Received event network-vif-deleted-da476e21-90c9-4f68-926c-883707c320ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.721800] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.721800] env[62235]: value = "task-1271564" [ 787.721800] env[62235]: _type = "Task" [ 787.721800] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.730896] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271564, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.790195] env[62235]: INFO nova.compute.manager [-] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Took 1.25 seconds to deallocate network for instance. [ 788.163497] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 590440de-4b3b-4019-b0c3-58b5ad329c88 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.231304] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271564, 'name': CreateVM_Task, 'duration_secs': 0.316351} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.232022] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 788.232022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.232146] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.232449] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 788.233077] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb2357a5-7c9e-4b38-a2f4-5f11fb75fa29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.237668] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 788.237668] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5270885f-c645-650f-f075-03d94d032424" [ 788.237668] env[62235]: _type = "Task" [ 788.237668] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.245339] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5270885f-c645-650f-f075-03d94d032424, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.297622] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.666658] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0e9182f8-81e1-4ca2-97b6-0066faba0b22 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.748058] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5270885f-c645-650f-f075-03d94d032424, 'name': SearchDatastore_Task, 'duration_secs': 0.009038} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.748382] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.748609] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.748835] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.748978] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.749167] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.749413] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fe475a2-d1a0-4bd2-8586-a2b0dfb33d59 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.757759] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.757898] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.758570] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-505e9888-54a2-45a4-b78e-656aa89461b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.763486] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 788.763486] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521b957b-424a-c0b8-8ca5-ae891e2cb639" [ 788.763486] env[62235]: _type = "Task" [ 788.763486] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.770615] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521b957b-424a-c0b8-8ca5-ae891e2cb639, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.169106] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance d165f549-1759-41bf-86fd-b77793bd49c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.274661] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521b957b-424a-c0b8-8ca5-ae891e2cb639, 'name': SearchDatastore_Task, 'duration_secs': 0.007996} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.276139] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32de5341-37af-4359-9c68-202a36c2fbbb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.280929] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 789.280929] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a6030f-10fc-a824-7f5f-62749f2748da" [ 789.280929] env[62235]: _type = "Task" [ 789.280929] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.288657] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a6030f-10fc-a824-7f5f-62749f2748da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.671924] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ced2aa67-ce75-4af5-9c55-773835d05733 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.790880] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a6030f-10fc-a824-7f5f-62749f2748da, 'name': SearchDatastore_Task, 'duration_secs': 0.011506} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.791207] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.791475] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 789.791741] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22c3db78-1f6b-455f-93a0-b82bef8fbbd0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.798363] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 789.798363] env[62235]: value = "task-1271565" [ 789.798363] env[62235]: _type = "Task" [ 789.798363] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.805420] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271565, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.175934] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b5a1da09-2137-4434-a1a4-174b196b61b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.308955] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271565, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.409533} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.309323] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 790.309577] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.309838] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b05dc809-99c1-4fdf-9e8d-2e65ec29765d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.316571] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 790.316571] env[62235]: value = "task-1271566" [ 790.316571] env[62235]: _type = "Task" [ 790.316571] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.325248] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271566, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.679223] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 876b7602-ccf0-4aed-90e2-d07fbf548fb6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.827321] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271566, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058162} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.827597] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.828331] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd082c91-e825-4c7d-a7af-04926920fd95 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.846787] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.847010] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b46c92a1-393f-4384-b84c-ee63fc66235b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.865876] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 790.865876] env[62235]: value = "task-1271567" [ 790.865876] env[62235]: _type = "Task" [ 790.865876] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.872897] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.184126] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 06ec6b9c-bef0-4f96-a185-4315961de7f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.376335] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271567, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.687082] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 80a803d1-c034-42b6-a15d-9fc3893953ad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.876972] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271567, 'name': ReconfigVM_Task, 'duration_secs': 0.847036} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.877289] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Reconfigured VM instance instance-00000039 to attach disk [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.877905] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-827786fb-6eaa-46e6-b85b-4ef8f3f81aca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.884901] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 791.884901] env[62235]: value = "task-1271568" [ 791.884901] env[62235]: _type = "Task" [ 791.884901] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.892105] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271568, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.190211] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 00825c9e-b398-4edc-a5ed-e4b5b476af1c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.395290] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271568, 'name': Rename_Task, 'duration_secs': 0.12359} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.395580] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 792.395829] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-296eaf9e-26d9-49d0-a8f3-2a915d1913a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.402466] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 792.402466] env[62235]: value = "task-1271569" [ 792.402466] env[62235]: _type = "Task" [ 792.402466] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.410075] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.694438] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.912669] env[62235]: DEBUG oslo_vmware.api [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271569, 'name': PowerOnVM_Task, 'duration_secs': 0.490389} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.912965] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.913182] env[62235]: INFO nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Took 5.27 seconds to spawn the instance on the hypervisor. [ 792.913370] env[62235]: DEBUG nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.914119] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d2987a-baf5-4a68-8f20-dc4be8c43c7f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.196529] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b126a2c9-62e9-49a1-91ae-7a703e5fa490 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.196857] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 793.196937] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 793.390464] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf53d25-2f5b-4176-90ab-abcd3a5984c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.397862] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17384905-41ae-49a2-abe0-1706ecfe5628 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.434036] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7373ad1b-0f52-4fb3-baae-cd5fcadabc89 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.436933] env[62235]: INFO nova.compute.manager [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Took 28.92 seconds to build instance. [ 793.443398] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215d9eb0-8db1-4bda-902c-868ba485d687 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.457202] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.939511] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4f0b779b-0f7c-4f4b-ade6-f8aee959d83f tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.070s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.963374] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.966607] env[62235]: INFO nova.compute.manager [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Rebuilding instance [ 794.003532] env[62235]: DEBUG nova.compute.manager [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.004405] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35778238-0b4a-4d75-bae7-cad91cab7f18 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.442925] env[62235]: DEBUG nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.470807] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 794.471133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.857s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.471377] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.893s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.473048] env[62235]: INFO nova.compute.claims [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.475930] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 794.476134] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Cleaning up deleted instances {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 794.515460] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 794.516014] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05121768-aa80-4ddd-ab54-7b92a4035bc5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.523589] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 794.523589] env[62235]: value = "task-1271570" [ 794.523589] env[62235]: _type = "Task" [ 794.523589] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.532318] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.963106] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.981812] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] There are 4 instances to clean {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 794.982083] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 430056ca-ccba-45c5-adf7-9c068f81f9ec] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 795.033550] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271570, 'name': PowerOffVM_Task, 'duration_secs': 0.179195} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.033815] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 795.034039] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.034756] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e7f287-e7d7-42d3-8d46-6deaf816000b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.040968] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 795.041201] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f9abda6-90f9-4868-9975-df7c26d88f78 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.068716] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 795.068716] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 795.068716] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleting the datastore file [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 795.068716] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b7b794e-ac06-4b41-8b6c-214a7c509e17 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.075064] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 795.075064] env[62235]: value = "task-1271572" [ 795.075064] env[62235]: _type = "Task" [ 795.075064] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.084085] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271572, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.487472] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: be4fb088-8b01-4245-82fb-b0e085825597] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 795.584267] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271572, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090664} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.586684] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 795.586878] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 795.587064] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 795.712023] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1984d541-6657-4af9-82c8-75111a936b14 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.719981] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ccc9d7-ee55-47b1-978e-ab26fe90fedb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.753108] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd25807f-f6a5-4870-bee0-7f04913ebbdc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.762231] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311c67cb-28fc-447f-a41a-e7c4c317bceb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.777949] env[62235]: DEBUG nova.compute.provider_tree [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.991507] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 4e20ae82-9a11-433b-ac4e-a85404cf4ec6] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 796.280586] env[62235]: DEBUG nova.scheduler.client.report [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.494557] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d1ed75d6-8461-4eb2-a2b4-2decb9eba5c0] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 796.615135] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.615421] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.615593] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.615777] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.615929] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.616094] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.616304] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.616462] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.616628] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.616791] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.616961] env[62235]: DEBUG nova.virt.hardware [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.617814] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250ec900-162c-4326-98d9-177df77d53b4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.625777] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c270b2-c7f2-4dca-9e82-72d07c4fcdbb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.638628] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Instance VIF info [] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.644052] env[62235]: DEBUG oslo.service.loopingcall [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.644288] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 796.644496] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-481fbbfc-cd8e-4e6c-9ea0-bbb98f28dff7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.660427] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.660427] env[62235]: value = "task-1271573" [ 796.660427] env[62235]: _type = "Task" [ 796.660427] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.667709] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271573, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.785251] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.785769] env[62235]: DEBUG nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 796.788490] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.980s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.788678] env[62235]: DEBUG nova.objects.instance [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 796.998482] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.998659] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Cleaning up deleted instances with incomplete migration {{(pid=62235) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 797.170881] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271573, 'name': CreateVM_Task, 'duration_secs': 0.302028} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.171069] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.171493] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.171651] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.171961] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.172224] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-799a4211-f1d2-431f-a0d8-5a1ff4c2741a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.176535] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 797.176535] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526a1677-f60c-8454-156a-a04efe9325d3" [ 797.176535] env[62235]: _type = "Task" [ 797.176535] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.184935] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526a1677-f60c-8454-156a-a04efe9325d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.293202] env[62235]: DEBUG nova.compute.utils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.294841] env[62235]: DEBUG nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.294841] env[62235]: DEBUG nova.network.neutron [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 797.358365] env[62235]: DEBUG nova.policy [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '738ee55c831c48f894efbcef71bc20a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c943a445f67f4021bad11723e0836537', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.500837] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.632159] env[62235]: DEBUG nova.network.neutron [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Successfully created port: eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.686648] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526a1677-f60c-8454-156a-a04efe9325d3, 'name': SearchDatastore_Task, 'duration_secs': 0.010063} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.686948] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.687218] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.687460] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.687607] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.687785] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.688052] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a601b3e-6151-4acb-9294-2ecd1b9cd3fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.700459] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.700640] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.701363] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf872441-c871-4bbc-8cf1-b38c2c94f10d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.707206] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 797.707206] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527a63a9-db4f-dab4-9253-d2ad6254e6c2" [ 797.707206] env[62235]: _type = "Task" [ 797.707206] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.714707] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527a63a9-db4f-dab4-9253-d2ad6254e6c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.801107] env[62235]: DEBUG nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.808486] env[62235]: DEBUG oslo_concurrency.lockutils [None req-dafd1cdd-92bf-4c00-a135-9cfb44907736 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.811880] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.317s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.811880] env[62235]: INFO nova.compute.claims [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.218549] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527a63a9-db4f-dab4-9253-d2ad6254e6c2, 'name': SearchDatastore_Task, 'duration_secs': 0.0165} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.219369] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ee2cc35-6fa3-4d72-b1f0-60bcdb13cfad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.224847] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 798.224847] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527bacff-9497-56ae-0b96-e176fb093a0d" [ 798.224847] env[62235]: _type = "Task" [ 798.224847] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.232435] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527bacff-9497-56ae-0b96-e176fb093a0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.735660] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527bacff-9497-56ae-0b96-e176fb093a0d, 'name': SearchDatastore_Task, 'duration_secs': 0.008743} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.736040] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.736186] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 798.736450] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91e8fc9d-06eb-443c-b1c2-1bfc6fc68461 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.743163] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 798.743163] env[62235]: value = "task-1271574" [ 798.743163] env[62235]: _type = "Task" [ 798.743163] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.750501] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271574, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.814962] env[62235]: DEBUG nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.845019] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.845019] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.845019] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.845225] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.845225] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.845225] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.845225] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.845225] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.845377] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.845377] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.845377] env[62235]: DEBUG nova.virt.hardware [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.846103] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e536f0-a494-4241-bcbb-50151ddd9959 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.853685] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93a2d1d-4d0a-4cd9-8487-76d56f7c00ee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.072041] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23bd4e4-6364-4ada-b5de-c87294768779 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.079909] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6faa5b-489a-4523-af81-16819ce2e1bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.113323] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc223da-05ef-488c-b62a-ab65dd9777a9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.123024] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829cb321-2163-45a5-97a2-56ca9e864754 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.129858] env[62235]: DEBUG nova.compute.manager [req-dd687323-052a-4a1c-9429-ff5c680acf66 req-917e17b7-93a9-4ec5-9280-93c957cd258c service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Received event network-vif-plugged-eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.130127] env[62235]: DEBUG oslo_concurrency.lockutils [req-dd687323-052a-4a1c-9429-ff5c680acf66 req-917e17b7-93a9-4ec5-9280-93c957cd258c service nova] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.130396] env[62235]: DEBUG oslo_concurrency.lockutils [req-dd687323-052a-4a1c-9429-ff5c680acf66 req-917e17b7-93a9-4ec5-9280-93c957cd258c service nova] Lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.130569] env[62235]: DEBUG oslo_concurrency.lockutils [req-dd687323-052a-4a1c-9429-ff5c680acf66 req-917e17b7-93a9-4ec5-9280-93c957cd258c service nova] Lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.130786] env[62235]: DEBUG nova.compute.manager [req-dd687323-052a-4a1c-9429-ff5c680acf66 req-917e17b7-93a9-4ec5-9280-93c957cd258c service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] No waiting events found dispatching network-vif-plugged-eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.131016] env[62235]: WARNING nova.compute.manager [req-dd687323-052a-4a1c-9429-ff5c680acf66 req-917e17b7-93a9-4ec5-9280-93c957cd258c service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Received unexpected event network-vif-plugged-eed03567-8c20-4e56-902c-fcd6bb25b56b for instance with vm_state building and task_state spawning. [ 799.142628] env[62235]: DEBUG nova.compute.provider_tree [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.204775] env[62235]: DEBUG nova.network.neutron [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Successfully updated port: eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.253093] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271574, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.647045] env[62235]: DEBUG nova.scheduler.client.report [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.708245] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.708245] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.708369] env[62235]: DEBUG nova.network.neutron [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.754651] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271574, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551037} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.754913] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.755137] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.755383] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4af4fc95-4740-450a-b26e-1f9def4a11d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.761576] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 799.761576] env[62235]: value = "task-1271575" [ 799.761576] env[62235]: _type = "Task" [ 799.761576] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.768371] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.151712] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.152279] env[62235]: DEBUG nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.154975] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.286s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.155205] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.157295] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.977s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.158711] env[62235]: INFO nova.compute.claims [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.180224] env[62235]: INFO nova.scheduler.client.report [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Deleted allocations for instance 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba [ 800.239119] env[62235]: DEBUG nova.network.neutron [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.271275] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097534} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.271535] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.272287] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b65911f-d415-4edb-ae60-24b656c40536 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.291890] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.294668] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3a9c1f4-11b9-4de9-8bf0-45fe59c6a4bb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.313932] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 800.313932] env[62235]: value = "task-1271576" [ 800.313932] env[62235]: _type = "Task" [ 800.313932] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.321684] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271576, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.412233] env[62235]: DEBUG nova.network.neutron [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.663513] env[62235]: DEBUG nova.compute.utils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.667065] env[62235]: DEBUG nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.667065] env[62235]: DEBUG nova.network.neutron [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.687289] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e00ef9da-dd5a-4845-817f-402236f1a361 tempest-ServerShowV257Test-1798414375 tempest-ServerShowV257Test-1798414375-project-member] Lock "8559690a-bdd3-4ec3-98e1-dfc78c6db8ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.566s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.709722] env[62235]: DEBUG nova.policy [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d7ef1bc26534632ad28710f9df9803c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e22686dbba4420dad98d5de1f4fd449', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 800.823528] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271576, 'name': ReconfigVM_Task, 'duration_secs': 0.2899} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.823813] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Reconfigured VM instance instance-00000039 to attach disk [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44/b4a3777f-2ee5-464b-84eb-0722d59f2b44.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.824405] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a25b9969-6dab-4a0e-aa89-059d71df5e62 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.830840] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 800.830840] env[62235]: value = "task-1271577" [ 800.830840] env[62235]: _type = "Task" [ 800.830840] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.838228] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271577, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.915191] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.915674] env[62235]: DEBUG nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Instance network_info: |[{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.916012] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:cc:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37fb1918-d178-4e12-93e6-316381e78be4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eed03567-8c20-4e56-902c-fcd6bb25b56b', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.923711] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Creating folder: Project (c943a445f67f4021bad11723e0836537). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.923979] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c3ff764-8fe1-491a-a994-fb92594ee351 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.935013] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Created folder: Project (c943a445f67f4021bad11723e0836537) in parent group-v273362. [ 800.935269] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Creating folder: Instances. Parent ref: group-v273404. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.935519] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3ab6d64-a9ee-4bd9-a2ac-ca7461b0b9e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.946261] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Created folder: Instances in parent group-v273404. [ 800.946519] env[62235]: DEBUG oslo.service.loopingcall [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.947153] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.947153] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa04ddb4-68bf-45c9-9c43-76c331027f3f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.966427] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.966427] env[62235]: value = "task-1271580" [ 800.966427] env[62235]: _type = "Task" [ 800.966427] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.971436] env[62235]: DEBUG nova.network.neutron [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Successfully created port: 19984205-bfbe-4c4a-a3e6-389954e0aa5a {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.978203] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271580, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.162328] env[62235]: DEBUG nova.compute.manager [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Received event network-changed-eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.162552] env[62235]: DEBUG nova.compute.manager [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Refreshing instance network info cache due to event network-changed-eed03567-8c20-4e56-902c-fcd6bb25b56b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 801.162744] env[62235]: DEBUG oslo_concurrency.lockutils [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.162886] env[62235]: DEBUG oslo_concurrency.lockutils [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.163161] env[62235]: DEBUG nova.network.neutron [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Refreshing network info cache for port eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.167342] env[62235]: DEBUG nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.342902] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271577, 'name': Rename_Task, 'duration_secs': 0.143552} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.344139] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 801.344139] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88e957f3-1ddf-4283-a225-8289f7acf7db {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.350185] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 801.350185] env[62235]: value = "task-1271581" [ 801.350185] env[62235]: _type = "Task" [ 801.350185] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.359929] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271581, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.430713] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b35e61-f438-4e1b-ae3a-c51d1a335ce8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.438550] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8e7690-099d-483d-b8a4-0fbf65705508 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.470248] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365ad346-2bda-43e6-bb3f-54c8661c3813 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.477434] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271580, 'name': CreateVM_Task, 'duration_secs': 0.30615} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.479307] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.480044] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.480327] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.480959] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.482107] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd9173f-ac4d-4c2c-a0af-969affb807e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.485816] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba75c1c5-71cb-421f-8254-e6ffb7fc8b90 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.500034] env[62235]: DEBUG nova.compute.provider_tree [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.502613] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 801.502613] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523d2412-a015-49a2-05ef-f5afa9b9f331" [ 801.502613] env[62235]: _type = "Task" [ 801.502613] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.511790] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523d2412-a015-49a2-05ef-f5afa9b9f331, 'name': SearchDatastore_Task, 'duration_secs': 0.008701} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.512276] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.512454] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.512684] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.512829] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.513020] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.513289] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-698f2112-82b8-48da-bddc-cefa454763e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.520675] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.520845] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.521552] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3f91285-f0c9-4f55-bf8c-d665aae3f9f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.527144] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 801.527144] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529a47c9-98f4-05ae-f5f2-3f7a7cfb375b" [ 801.527144] env[62235]: _type = "Task" [ 801.527144] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.535933] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529a47c9-98f4-05ae-f5f2-3f7a7cfb375b, 'name': SearchDatastore_Task, 'duration_secs': 0.007338} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.536695] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58730a45-3782-404a-b2c7-21fce82427a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.541334] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 801.541334] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522f674b-7ac1-7711-4410-db7a684d308e" [ 801.541334] env[62235]: _type = "Task" [ 801.541334] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.549273] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522f674b-7ac1-7711-4410-db7a684d308e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.862061] env[62235]: DEBUG oslo_vmware.api [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271581, 'name': PowerOnVM_Task, 'duration_secs': 0.406803} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.862352] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.862499] env[62235]: DEBUG nova.compute.manager [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.863307] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d2420a-079c-450c-adbd-b31231eb7c4f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.866596] env[62235]: DEBUG nova.network.neutron [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updated VIF entry in instance network info cache for port eed03567-8c20-4e56-902c-fcd6bb25b56b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.866908] env[62235]: DEBUG nova.network.neutron [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.004164] env[62235]: DEBUG nova.scheduler.client.report [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.052663] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522f674b-7ac1-7711-4410-db7a684d308e, 'name': SearchDatastore_Task, 'duration_secs': 0.009118} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.052934] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.053254] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 3ea7db36-ebde-4523-b803-ad121a00269b/3ea7db36-ebde-4523-b803-ad121a00269b.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.053530] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b4fc3f2-335b-4a25-b596-6a4e256462d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.061156] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 802.061156] env[62235]: value = "task-1271582" [ 802.061156] env[62235]: _type = "Task" [ 802.061156] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.069912] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.182724] env[62235]: DEBUG nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.206423] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.206692] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.206853] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.207054] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.207213] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.207365] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.207577] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.207735] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.207902] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.208074] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.208254] env[62235]: DEBUG nova.virt.hardware [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.209087] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8574ff5c-5d91-458a-9b09-1ea324ff7c3e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.217571] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd5d321-c9c6-4802-ba33-a8f67138565f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.373297] env[62235]: DEBUG oslo_concurrency.lockutils [req-76c455a8-1319-4cd8-85a7-8ef25a46a912 req-38aa97d1-b569-4715-90b5-0c235d6dcfb7 service nova] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.379618] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.512025] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.512025] env[62235]: DEBUG nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.512991] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.132s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.514938] env[62235]: INFO nova.compute.claims [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.572585] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.648984] env[62235]: DEBUG nova.network.neutron [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Successfully updated port: 19984205-bfbe-4c4a-a3e6-389954e0aa5a {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 802.872650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.873264] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.873571] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.874105] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.874227] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.877208] env[62235]: INFO nova.compute.manager [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Terminating instance [ 802.881181] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "refresh_cache-b4a3777f-2ee5-464b-84eb-0722d59f2b44" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.881409] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "refresh_cache-b4a3777f-2ee5-464b-84eb-0722d59f2b44" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.881621] env[62235]: DEBUG nova.network.neutron [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.022530] env[62235]: DEBUG nova.compute.utils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.024666] env[62235]: DEBUG nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.024852] env[62235]: DEBUG nova.network.neutron [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.061076] env[62235]: DEBUG nova.policy [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53ee098b2f8140c8a73ff4827406236f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c23977a4e494801a8dcb2e6a7c52137', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.072379] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.922208} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.072644] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 3ea7db36-ebde-4523-b803-ad121a00269b/3ea7db36-ebde-4523-b803-ad121a00269b.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.072861] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.073136] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7654c503-e0a4-4b2a-9b66-6937b8da9f25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.079383] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 803.079383] env[62235]: value = "task-1271583" [ 803.079383] env[62235]: _type = "Task" [ 803.079383] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.086835] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271583, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.154244] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "refresh_cache-63e16fd4-3a5a-4231-9013-c6a01f0eab80" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.154244] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "refresh_cache-63e16fd4-3a5a-4231-9013-c6a01f0eab80" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.154244] env[62235]: DEBUG nova.network.neutron [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.192453] env[62235]: DEBUG nova.compute.manager [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Received event network-vif-plugged-19984205-bfbe-4c4a-a3e6-389954e0aa5a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.192681] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] Acquiring lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.192889] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.193072] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.193242] env[62235]: DEBUG nova.compute.manager [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] No waiting events found dispatching network-vif-plugged-19984205-bfbe-4c4a-a3e6-389954e0aa5a {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.193496] env[62235]: WARNING nova.compute.manager [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Received unexpected event network-vif-plugged-19984205-bfbe-4c4a-a3e6-389954e0aa5a for instance with vm_state building and task_state spawning. [ 803.193685] env[62235]: DEBUG nova.compute.manager [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Received event network-changed-19984205-bfbe-4c4a-a3e6-389954e0aa5a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.193841] env[62235]: DEBUG nova.compute.manager [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Refreshing instance network info cache due to event network-changed-19984205-bfbe-4c4a-a3e6-389954e0aa5a. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 803.194055] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] Acquiring lock "refresh_cache-63e16fd4-3a5a-4231-9013-c6a01f0eab80" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.345840] env[62235]: DEBUG nova.network.neutron [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Successfully created port: fa915021-a55c-4801-90c1-4c80e3d88c70 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.409903] env[62235]: DEBUG nova.network.neutron [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.506523] env[62235]: DEBUG nova.network.neutron [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.525679] env[62235]: DEBUG nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.591679] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271583, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062408} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.591945] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.592719] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ed6a4c-7175-4ff1-9544-091653c302a1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.618118] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 3ea7db36-ebde-4523-b803-ad121a00269b/3ea7db36-ebde-4523-b803-ad121a00269b.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.618118] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-027ef4cf-6534-466b-a234-83e337b7e635 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.640202] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 803.640202] env[62235]: value = "task-1271584" [ 803.640202] env[62235]: _type = "Task" [ 803.640202] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.651395] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271584, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.702620] env[62235]: DEBUG nova.network.neutron [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.816538] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd5b904-0d38-4111-9b0d-2e88108bc1a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.824717] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637bce60-56c2-4936-b684-39ce3c4217d0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.855113] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863cc065-d3d0-4daf-a337-1394fd552a54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.858324] env[62235]: DEBUG nova.network.neutron [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Updating instance_info_cache with network_info: [{"id": "19984205-bfbe-4c4a-a3e6-389954e0aa5a", "address": "fa:16:3e:31:72:52", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19984205-bf", "ovs_interfaceid": "19984205-bfbe-4c4a-a3e6-389954e0aa5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.865046] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96fa20d-8952-4154-967c-9ee7986f6566 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.880721] env[62235]: DEBUG nova.compute.provider_tree [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.009546] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "refresh_cache-b4a3777f-2ee5-464b-84eb-0722d59f2b44" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.009982] env[62235]: DEBUG nova.compute.manager [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.010200] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.011071] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be1cdab-68b7-43d0-8aea-9c6a1d2af5ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.019352] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.019594] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bf487e7-910e-418c-ab6e-3f44cb4f06be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.025678] env[62235]: DEBUG oslo_vmware.api [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 804.025678] env[62235]: value = "task-1271585" [ 804.025678] env[62235]: _type = "Task" [ 804.025678] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.037718] env[62235]: DEBUG oslo_vmware.api [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.151607] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271584, 'name': ReconfigVM_Task, 'duration_secs': 0.257659} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.151921] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 3ea7db36-ebde-4523-b803-ad121a00269b/3ea7db36-ebde-4523-b803-ad121a00269b.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.152546] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6a24b1d-97c0-4fd5-a876-e1e5cf64911e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.158175] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 804.158175] env[62235]: value = "task-1271586" [ 804.158175] env[62235]: _type = "Task" [ 804.158175] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.166265] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271586, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.361400] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "refresh_cache-63e16fd4-3a5a-4231-9013-c6a01f0eab80" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.361750] env[62235]: DEBUG nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Instance network_info: |[{"id": "19984205-bfbe-4c4a-a3e6-389954e0aa5a", "address": "fa:16:3e:31:72:52", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19984205-bf", "ovs_interfaceid": "19984205-bfbe-4c4a-a3e6-389954e0aa5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.362060] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] Acquired lock "refresh_cache-63e16fd4-3a5a-4231-9013-c6a01f0eab80" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.362248] env[62235]: DEBUG nova.network.neutron [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Refreshing network info cache for port 19984205-bfbe-4c4a-a3e6-389954e0aa5a {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.363449] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:72:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19984205-bfbe-4c4a-a3e6-389954e0aa5a', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.371998] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating folder: Project (1e22686dbba4420dad98d5de1f4fd449). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.374904] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc0a09b7-caf3-40d3-aa39-9c9e04ac82c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.383677] env[62235]: DEBUG nova.scheduler.client.report [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.389688] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Created folder: Project (1e22686dbba4420dad98d5de1f4fd449) in parent group-v273362. [ 804.389688] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating folder: Instances. Parent ref: group-v273407. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.389688] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16fcc623-baa0-4e2b-b915-f27b74f6aeed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.397856] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Created folder: Instances in parent group-v273407. [ 804.398096] env[62235]: DEBUG oslo.service.loopingcall [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.398750] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 804.398984] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f31d58a-87f0-4597-b468-34ec1918ee1f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.419513] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.419513] env[62235]: value = "task-1271589" [ 804.419513] env[62235]: _type = "Task" [ 804.419513] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.427206] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271589, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.536404] env[62235]: DEBUG oslo_vmware.api [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271585, 'name': PowerOffVM_Task, 'duration_secs': 0.120514} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.536768] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.537043] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.537350] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1bb6746-f199-4b51-abe1-0030ab547bd0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.539816] env[62235]: DEBUG nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.565834] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 804.566073] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 804.566295] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleting the datastore file [datastore2] b4a3777f-2ee5-464b-84eb-0722d59f2b44 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.566566] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d7cf957-ade7-4be2-82e6-8bbac54679c7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.570423] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.570686] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.570877] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.571106] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.571287] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.571468] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.571707] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.571908] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.572230] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.572448] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.572717] env[62235]: DEBUG nova.virt.hardware [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.573504] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1be5074-a3c9-467a-acf3-4ba500c126dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.582023] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b8569a-69b0-4c14-ad5d-06729f2d868e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.586222] env[62235]: DEBUG oslo_vmware.api [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 804.586222] env[62235]: value = "task-1271591" [ 804.586222] env[62235]: _type = "Task" [ 804.586222] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.604233] env[62235]: DEBUG oslo_vmware.api [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.667311] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271586, 'name': Rename_Task, 'duration_secs': 0.133548} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.667634] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.667894] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0268e159-82f7-4e0e-aa0d-7714060722c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.675457] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 804.675457] env[62235]: value = "task-1271592" [ 804.675457] env[62235]: _type = "Task" [ 804.675457] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.683187] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271592, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.888067] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.888639] env[62235]: DEBUG nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.891479] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.242s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.893108] env[62235]: INFO nova.compute.claims [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.931168] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271589, 'name': CreateVM_Task, 'duration_secs': 0.320472} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.931168] env[62235]: DEBUG nova.network.neutron [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Updated VIF entry in instance network info cache for port 19984205-bfbe-4c4a-a3e6-389954e0aa5a. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.931265] env[62235]: DEBUG nova.network.neutron [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Updating instance_info_cache with network_info: [{"id": "19984205-bfbe-4c4a-a3e6-389954e0aa5a", "address": "fa:16:3e:31:72:52", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19984205-bf", "ovs_interfaceid": "19984205-bfbe-4c4a-a3e6-389954e0aa5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.931724] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.933691] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.933982] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.934309] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.934998] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8123fec6-1f56-4397-8d71-3289698752f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.939593] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 804.939593] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52daa2fe-109d-7873-9418-4d2500eb12ff" [ 804.939593] env[62235]: _type = "Task" [ 804.939593] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.949760] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52daa2fe-109d-7873-9418-4d2500eb12ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.069867] env[62235]: DEBUG nova.network.neutron [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Successfully updated port: fa915021-a55c-4801-90c1-4c80e3d88c70 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.097238] env[62235]: DEBUG oslo_vmware.api [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271591, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093151} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.097536] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.097742] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 805.097921] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.098115] env[62235]: INFO nova.compute.manager [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Took 1.09 seconds to destroy the instance on the hypervisor. [ 805.098493] env[62235]: DEBUG oslo.service.loopingcall [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.098696] env[62235]: DEBUG nova.compute.manager [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 805.098790] env[62235]: DEBUG nova.network.neutron [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 805.123650] env[62235]: DEBUG nova.network.neutron [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.185476] env[62235]: DEBUG oslo_vmware.api [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271592, 'name': PowerOnVM_Task, 'duration_secs': 0.45068} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.185747] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.185945] env[62235]: INFO nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Took 6.37 seconds to spawn the instance on the hypervisor. [ 805.186139] env[62235]: DEBUG nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.186897] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0daf5732-7017-4fba-99be-2b54a084dd6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.222432] env[62235]: DEBUG nova.compute.manager [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Received event network-vif-plugged-fa915021-a55c-4801-90c1-4c80e3d88c70 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.222707] env[62235]: DEBUG oslo_concurrency.lockutils [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] Acquiring lock "590440de-4b3b-4019-b0c3-58b5ad329c88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.222929] env[62235]: DEBUG oslo_concurrency.lockutils [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.223137] env[62235]: DEBUG oslo_concurrency.lockutils [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.223320] env[62235]: DEBUG nova.compute.manager [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] No waiting events found dispatching network-vif-plugged-fa915021-a55c-4801-90c1-4c80e3d88c70 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 805.223489] env[62235]: WARNING nova.compute.manager [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Received unexpected event network-vif-plugged-fa915021-a55c-4801-90c1-4c80e3d88c70 for instance with vm_state building and task_state spawning. [ 805.223650] env[62235]: DEBUG nova.compute.manager [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Received event network-changed-fa915021-a55c-4801-90c1-4c80e3d88c70 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.223802] env[62235]: DEBUG nova.compute.manager [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Refreshing instance network info cache due to event network-changed-fa915021-a55c-4801-90c1-4c80e3d88c70. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.223987] env[62235]: DEBUG oslo_concurrency.lockutils [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] Acquiring lock "refresh_cache-590440de-4b3b-4019-b0c3-58b5ad329c88" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.224138] env[62235]: DEBUG oslo_concurrency.lockutils [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] Acquired lock "refresh_cache-590440de-4b3b-4019-b0c3-58b5ad329c88" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.224633] env[62235]: DEBUG nova.network.neutron [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Refreshing network info cache for port fa915021-a55c-4801-90c1-4c80e3d88c70 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.398249] env[62235]: DEBUG nova.compute.utils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.401709] env[62235]: DEBUG nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.401709] env[62235]: DEBUG nova.network.neutron [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.434204] env[62235]: DEBUG oslo_concurrency.lockutils [req-ef8d40e8-1f8c-4766-babc-c2a33cb71d3a req-eab5b200-6de4-4e33-95bd-e3009acbe77b service nova] Releasing lock "refresh_cache-63e16fd4-3a5a-4231-9013-c6a01f0eab80" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.439591] env[62235]: DEBUG nova.policy [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6018b2929a594c358fe543f4fa2d2f46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6832a93601f84138aeed5a906f24394b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.450649] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52daa2fe-109d-7873-9418-4d2500eb12ff, 'name': SearchDatastore_Task, 'duration_secs': 0.00869} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.450929] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.451166] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.451401] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.451583] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.451718] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.451960] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f55fcdab-76c3-484b-a0bf-acd408ab549b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.463480] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.463657] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.464372] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f183fad-449b-4754-8599-ba81f230d17c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.469276] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 805.469276] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525cb813-3d87-8a61-acb9-d111013c24da" [ 805.469276] env[62235]: _type = "Task" [ 805.469276] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.476522] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525cb813-3d87-8a61-acb9-d111013c24da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.571478] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "refresh_cache-590440de-4b3b-4019-b0c3-58b5ad329c88" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.627275] env[62235]: DEBUG nova.network.neutron [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.705030] env[62235]: INFO nova.compute.manager [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Took 39.14 seconds to build instance. [ 805.756424] env[62235]: DEBUG nova.network.neutron [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Successfully created port: f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.780111] env[62235]: DEBUG nova.network.neutron [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.901242] env[62235]: DEBUG nova.network.neutron [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.902694] env[62235]: DEBUG nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.984929] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525cb813-3d87-8a61-acb9-d111013c24da, 'name': SearchDatastore_Task, 'duration_secs': 0.029889} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.986135] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45ef4ad2-d9a9-4f38-979b-1bcbba782e04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.993478] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 805.993478] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520d5311-04b2-e4aa-67aa-5022ad629fec" [ 805.993478] env[62235]: _type = "Task" [ 805.993478] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.001219] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520d5311-04b2-e4aa-67aa-5022ad629fec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.129971] env[62235]: INFO nova.compute.manager [-] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Took 1.03 seconds to deallocate network for instance. [ 806.190263] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80772a1-318b-4eef-a053-b7b1f984283d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.198333] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef43429-a41d-4b87-a555-b8814928125f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.229217] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a83968a-fd2f-41a3-9c45-0e302aed9a05 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.975s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.231255] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a00907a-71f3-4a98-9e36-dab2f3c04619 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.239564] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3508db-44d4-47f8-b156-249493d3795d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.253824] env[62235]: DEBUG nova.compute.provider_tree [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.406306] env[62235]: DEBUG oslo_concurrency.lockutils [req-a4273a24-b1f8-4437-b003-8b33b9ce6fb8 req-d49bbc23-f888-4f06-9597-d939563c9122 service nova] Releasing lock "refresh_cache-590440de-4b3b-4019-b0c3-58b5ad329c88" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.406702] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquired lock "refresh_cache-590440de-4b3b-4019-b0c3-58b5ad329c88" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.406907] env[62235]: DEBUG nova.network.neutron [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.507021] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520d5311-04b2-e4aa-67aa-5022ad629fec, 'name': SearchDatastore_Task, 'duration_secs': 0.008753} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.507021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.507021] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 63e16fd4-3a5a-4231-9013-c6a01f0eab80/63e16fd4-3a5a-4231-9013-c6a01f0eab80.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 806.507021] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4de503e9-ffde-4797-92df-fae2c8ff14ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.514153] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 806.514153] env[62235]: value = "task-1271593" [ 806.514153] env[62235]: _type = "Task" [ 806.514153] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.525786] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271593, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.636815] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.736720] env[62235]: DEBUG nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 806.759025] env[62235]: DEBUG nova.scheduler.client.report [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.916033] env[62235]: DEBUG nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.949666] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.953189] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.953189] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.953189] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.953189] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.953189] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.953444] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.953444] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.953444] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.953444] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.953444] env[62235]: DEBUG nova.virt.hardware [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.954595] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c00fa7-7735-42f4-ad80-9a589fa2bfeb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.959949] env[62235]: DEBUG nova.network.neutron [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.969154] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b533158-666b-411c-a1ac-7503089c7ba1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.030771] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271593, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438775} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.030771] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 63e16fd4-3a5a-4231-9013-c6a01f0eab80/63e16fd4-3a5a-4231-9013-c6a01f0eab80.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.030771] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.031060] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3be74a90-c560-4161-89df-c14f58ce537a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.037317] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 807.037317] env[62235]: value = "task-1271594" [ 807.037317] env[62235]: _type = "Task" [ 807.037317] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.045339] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271594, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.261058] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.262943] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.263472] env[62235]: DEBUG nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.266168] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.616s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.267809] env[62235]: INFO nova.compute.claims [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.456349] env[62235]: DEBUG nova.network.neutron [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Updating instance_info_cache with network_info: [{"id": "fa915021-a55c-4801-90c1-4c80e3d88c70", "address": "fa:16:3e:36:1c:17", "network": {"id": "5dfab083-8f66-4adf-818c-341ad59efa9d", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-587115759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c23977a4e494801a8dcb2e6a7c52137", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa915021-a5", "ovs_interfaceid": "fa915021-a55c-4801-90c1-4c80e3d88c70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.539976] env[62235]: DEBUG nova.compute.manager [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Received event network-changed-eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.540064] env[62235]: DEBUG nova.compute.manager [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Refreshing instance network info cache due to event network-changed-eed03567-8c20-4e56-902c-fcd6bb25b56b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 807.540451] env[62235]: DEBUG oslo_concurrency.lockutils [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.540682] env[62235]: DEBUG oslo_concurrency.lockutils [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.540886] env[62235]: DEBUG nova.network.neutron [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Refreshing network info cache for port eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 807.551230] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271594, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065681} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.552086] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.553057] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d7fe23-3cf4-408e-bf6f-ab5020e544a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.578020] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 63e16fd4-3a5a-4231-9013-c6a01f0eab80/63e16fd4-3a5a-4231-9013-c6a01f0eab80.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.578020] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e76c898-0bab-4179-bca6-4d60362b3c23 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.597080] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 807.597080] env[62235]: value = "task-1271595" [ 807.597080] env[62235]: _type = "Task" [ 807.597080] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.606626] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271595, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.673656] env[62235]: DEBUG nova.network.neutron [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Successfully updated port: f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.772083] env[62235]: DEBUG nova.compute.utils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.776591] env[62235]: DEBUG nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.776591] env[62235]: DEBUG nova.network.neutron [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 807.827171] env[62235]: DEBUG nova.policy [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.960713] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Releasing lock "refresh_cache-590440de-4b3b-4019-b0c3-58b5ad329c88" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.961278] env[62235]: DEBUG nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Instance network_info: |[{"id": "fa915021-a55c-4801-90c1-4c80e3d88c70", "address": "fa:16:3e:36:1c:17", "network": {"id": "5dfab083-8f66-4adf-818c-341ad59efa9d", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-587115759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c23977a4e494801a8dcb2e6a7c52137", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa915021-a5", "ovs_interfaceid": "fa915021-a55c-4801-90c1-4c80e3d88c70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.961542] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:1c:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd646f9d5-d2ad-4c22-bea5-85a965334de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa915021-a55c-4801-90c1-4c80e3d88c70', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.969731] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Creating folder: Project (3c23977a4e494801a8dcb2e6a7c52137). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.969953] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-904e722e-ab5a-409a-95c7-c2d5621d0826 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.979675] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Created folder: Project (3c23977a4e494801a8dcb2e6a7c52137) in parent group-v273362. [ 807.979846] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Creating folder: Instances. Parent ref: group-v273410. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.982120] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a597a0bd-ef0d-42c7-bd55-aa9976c93199 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.989047] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Created folder: Instances in parent group-v273410. [ 807.989047] env[62235]: DEBUG oslo.service.loopingcall [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.989185] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.989282] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b05e7d4-2832-437c-8c5e-ab0601501f43 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.009669] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.009669] env[62235]: value = "task-1271598" [ 808.009669] env[62235]: _type = "Task" [ 808.009669] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.017630] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271598, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.107891] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271595, 'name': ReconfigVM_Task, 'duration_secs': 0.27029} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.108227] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 63e16fd4-3a5a-4231-9013-c6a01f0eab80/63e16fd4-3a5a-4231-9013-c6a01f0eab80.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.109258] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3234cba-7a40-4ebb-8309-35f057045ace {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.114973] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 808.114973] env[62235]: value = "task-1271599" [ 808.114973] env[62235]: _type = "Task" [ 808.114973] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.125667] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271599, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.179508] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.179751] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.180210] env[62235]: DEBUG nova.network.neutron [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.280054] env[62235]: DEBUG nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.304884] env[62235]: DEBUG nova.network.neutron [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Successfully created port: 235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.446245] env[62235]: DEBUG nova.network.neutron [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updated VIF entry in instance network info cache for port eed03567-8c20-4e56-902c-fcd6bb25b56b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 808.446673] env[62235]: DEBUG nova.network.neutron [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.520886] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271598, 'name': CreateVM_Task, 'duration_secs': 0.329229} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.521270] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.522326] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.522510] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.522817] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.523081] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1abb2f56-083e-40c0-9d2d-80e414b10727 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.530987] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 808.530987] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fbe87a-ca1a-814d-718a-0aa2e495a1e8" [ 808.530987] env[62235]: _type = "Task" [ 808.530987] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.539524] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fbe87a-ca1a-814d-718a-0aa2e495a1e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.562330] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fe8f66-0049-423b-88da-dfcdccde149e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.570372] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97840020-f95a-46cf-9186-cc9e57420105 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.608084] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d246905-601a-4563-8dbb-5cebb41c392a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.615145] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52ca715-08d4-4af5-a113-7c270d586d4e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.631507] env[62235]: DEBUG nova.compute.provider_tree [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.635771] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271599, 'name': Rename_Task, 'duration_secs': 0.192646} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.636947] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 808.636947] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9cca9f46-7548-4030-a349-0939bc559d3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.644798] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 808.644798] env[62235]: value = "task-1271600" [ 808.644798] env[62235]: _type = "Task" [ 808.644798] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.656035] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271600, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.713506] env[62235]: DEBUG nova.network.neutron [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.849660] env[62235]: DEBUG nova.network.neutron [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.949348] env[62235]: DEBUG oslo_concurrency.lockutils [req-a5c5f21b-8af1-47c9-80d3-701a8e04ba30 req-7d3d5b61-64c9-4feb-8014-936179e77eab service nova] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.042438] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fbe87a-ca1a-814d-718a-0aa2e495a1e8, 'name': SearchDatastore_Task, 'duration_secs': 0.009084} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.042778] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.043015] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.043282] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.043447] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.043631] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.043895] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37780531-3bab-44b8-9d42-76d0f112cf3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.052241] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.052375] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.053074] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-248d3cdc-ace6-4dca-8ff1-5cd039180c8c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.057884] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 809.057884] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528396a5-7a7a-6dfc-bfe6-e3488a9d68cc" [ 809.057884] env[62235]: _type = "Task" [ 809.057884] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.065138] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528396a5-7a7a-6dfc-bfe6-e3488a9d68cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.137861] env[62235]: DEBUG nova.scheduler.client.report [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.155868] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271600, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.291970] env[62235]: DEBUG nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.316305] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.316670] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.316835] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.317029] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.317183] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.317328] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.317532] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.317705] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.317892] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.318070] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.318245] env[62235]: DEBUG nova.virt.hardware [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.319087] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49a0c36-03b3-4945-9d9f-1f97456683a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.328445] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10f55a6-66ff-4d14-9464-ccb62a6ad7a3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.352680] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.352976] env[62235]: DEBUG nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Instance network_info: |[{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.353353] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:60:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.360715] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating folder: Project (6832a93601f84138aeed5a906f24394b). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.360947] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f5414fe1-6023-4e60-8cc4-e43ca29af371 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.371502] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created folder: Project (6832a93601f84138aeed5a906f24394b) in parent group-v273362. [ 809.371681] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating folder: Instances. Parent ref: group-v273413. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.371891] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7e64112-4927-4307-8acd-00169ce132f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.381264] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created folder: Instances in parent group-v273413. [ 809.381482] env[62235]: DEBUG oslo.service.loopingcall [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.381659] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.381844] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5009778c-3f0b-4e22-a651-1ccb64e7a99c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.399849] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.399849] env[62235]: value = "task-1271603" [ 809.399849] env[62235]: _type = "Task" [ 809.399849] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.406726] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271603, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.570617] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528396a5-7a7a-6dfc-bfe6-e3488a9d68cc, 'name': SearchDatastore_Task, 'duration_secs': 0.00776} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.573339] env[62235]: DEBUG nova.compute.manager [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-vif-plugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.573566] env[62235]: DEBUG oslo_concurrency.lockutils [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.573777] env[62235]: DEBUG oslo_concurrency.lockutils [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.573943] env[62235]: DEBUG oslo_concurrency.lockutils [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.574177] env[62235]: DEBUG nova.compute.manager [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] No waiting events found dispatching network-vif-plugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 809.575025] env[62235]: WARNING nova.compute.manager [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received unexpected event network-vif-plugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 for instance with vm_state building and task_state spawning. [ 809.575025] env[62235]: DEBUG nova.compute.manager [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.575025] env[62235]: DEBUG nova.compute.manager [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing instance network info cache due to event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.575025] env[62235]: DEBUG oslo_concurrency.lockutils [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.575025] env[62235]: DEBUG oslo_concurrency.lockutils [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.575521] env[62235]: DEBUG nova.network.neutron [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.576284] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33f01523-497a-40d1-b4cb-d4039a1c3295 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.582365] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 809.582365] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fb9d34-fb19-acfe-a9c2-f0967a2b8d4e" [ 809.582365] env[62235]: _type = "Task" [ 809.582365] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.591262] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fb9d34-fb19-acfe-a9c2-f0967a2b8d4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.643222] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.643916] env[62235]: DEBUG nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.646838] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.889s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.648820] env[62235]: INFO nova.compute.claims [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.662033] env[62235]: DEBUG oslo_vmware.api [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271600, 'name': PowerOnVM_Task, 'duration_secs': 0.625219} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.662356] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 809.662623] env[62235]: INFO nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Took 7.48 seconds to spawn the instance on the hypervisor. [ 809.662822] env[62235]: DEBUG nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.663939] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525fb6a3-5adb-4869-936f-635e25581485 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.911241] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271603, 'name': CreateVM_Task, 'duration_secs': 0.349089} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.911424] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.912080] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.912251] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.912579] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.912832] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4fc8622-47e9-494c-af81-ef0bc4fbefe1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.917502] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 809.917502] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524c4416-46c6-6f0e-fb99-b25d5558befd" [ 809.917502] env[62235]: _type = "Task" [ 809.917502] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.925263] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524c4416-46c6-6f0e-fb99-b25d5558befd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.950291] env[62235]: DEBUG nova.network.neutron [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Successfully updated port: 235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 810.092488] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fb9d34-fb19-acfe-a9c2-f0967a2b8d4e, 'name': SearchDatastore_Task, 'duration_secs': 0.011533} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.092754] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.093012] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 590440de-4b3b-4019-b0c3-58b5ad329c88/590440de-4b3b-4019-b0c3-58b5ad329c88.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.093297] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-134b9a29-9146-4f57-8608-8893c5defbd0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.099546] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 810.099546] env[62235]: value = "task-1271604" [ 810.099546] env[62235]: _type = "Task" [ 810.099546] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.107606] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.154627] env[62235]: DEBUG nova.compute.utils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.156124] env[62235]: DEBUG nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.156313] env[62235]: DEBUG nova.network.neutron [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.184743] env[62235]: INFO nova.compute.manager [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Took 41.71 seconds to build instance. [ 810.245965] env[62235]: DEBUG nova.policy [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3ffdf6381df4aca8b4c613f13219262', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4456e079eae64f41b1596821d41bd275', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.430372] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524c4416-46c6-6f0e-fb99-b25d5558befd, 'name': SearchDatastore_Task, 'duration_secs': 0.008768} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.430787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.431034] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.431278] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.431428] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.431710] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.432152] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf326c24-49d4-4eaa-ba62-c0435d92cd9c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.435022] env[62235]: DEBUG nova.network.neutron [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updated VIF entry in instance network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 810.435475] env[62235]: DEBUG nova.network.neutron [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.448458] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.448674] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.449413] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78819ade-5ddb-4f3d-9d71-ce8490c5b963 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.452488] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.452624] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.452772] env[62235]: DEBUG nova.network.neutron [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 810.457854] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 810.457854] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e9da1c-d2f4-382f-b530-2fe63fe2d721" [ 810.457854] env[62235]: _type = "Task" [ 810.457854] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.468556] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e9da1c-d2f4-382f-b530-2fe63fe2d721, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.621788] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271604, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.433187} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.621849] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 590440de-4b3b-4019-b0c3-58b5ad329c88/590440de-4b3b-4019-b0c3-58b5ad329c88.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 810.622131] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.622640] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-259351ea-9596-403b-adba-aec6b851e5dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.630596] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 810.630596] env[62235]: value = "task-1271605" [ 810.630596] env[62235]: _type = "Task" [ 810.630596] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.641443] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.642523] env[62235]: DEBUG nova.network.neutron [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Successfully created port: aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.659895] env[62235]: DEBUG nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.686230] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daf70a66-4779-4d57-8b31-924b346abb19 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.772s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.896026] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9e73e2-3a46-4c26-8fc7-027d237da162 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.904728] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c31a82-c7d3-428b-af53-808985df7d9b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.937794] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d398f07-35f1-4f0d-a07b-ed0bdfac91fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.940815] env[62235]: DEBUG oslo_concurrency.lockutils [req-8c8f2f03-f0b8-46f9-898d-2991268543e7 req-8d6fdea1-2c1b-4e88-8865-f657bb6ef2de service nova] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.946094] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ad7048-6906-46b6-a2da-dddbb2246a84 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.964170] env[62235]: DEBUG nova.compute.provider_tree [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.977692] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e9da1c-d2f4-382f-b530-2fe63fe2d721, 'name': SearchDatastore_Task, 'duration_secs': 0.052184} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.979335] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09292e8c-b00e-4ca7-85a2-d13e1bafc127 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.986432] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 810.986432] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ed9611-7a66-fc70-a4e1-222f6085ee1b" [ 810.986432] env[62235]: _type = "Task" [ 810.986432] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.995510] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ed9611-7a66-fc70-a4e1-222f6085ee1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.016635] env[62235]: DEBUG nova.network.neutron [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.142972] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271605, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062203} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.143262] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.144055] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8fc98c-03cd-41df-a2ec-ea42be277252 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.167723] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 590440de-4b3b-4019-b0c3-58b5ad329c88/590440de-4b3b-4019-b0c3-58b5ad329c88.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.171695] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3156207-2798-4da6-bb96-efebcb6a2011 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.188451] env[62235]: DEBUG nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.196337] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 811.196337] env[62235]: value = "task-1271606" [ 811.196337] env[62235]: _type = "Task" [ 811.196337] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.196337] env[62235]: DEBUG nova.network.neutron [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.207321] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271606, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.468027] env[62235]: DEBUG nova.scheduler.client.report [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.496358] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ed9611-7a66-fc70-a4e1-222f6085ee1b, 'name': SearchDatastore_Task, 'duration_secs': 0.008906} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.496628] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.496885] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.497158] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57baf94a-25b0-4335-9cbf-73297dd13b71 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.503226] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 811.503226] env[62235]: value = "task-1271607" [ 811.503226] env[62235]: _type = "Task" [ 811.503226] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.510559] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271607, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.600692] env[62235]: DEBUG nova.compute.manager [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-vif-plugged-235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.600908] env[62235]: DEBUG oslo_concurrency.lockutils [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.601339] env[62235]: DEBUG oslo_concurrency.lockutils [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.601612] env[62235]: DEBUG oslo_concurrency.lockutils [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.601932] env[62235]: DEBUG nova.compute.manager [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] No waiting events found dispatching network-vif-plugged-235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 811.602144] env[62235]: WARNING nova.compute.manager [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received unexpected event network-vif-plugged-235e1a30-ec5f-4d9a-8085-8994ac4c8314 for instance with vm_state building and task_state spawning. [ 811.602344] env[62235]: DEBUG nova.compute.manager [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-changed-235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.602540] env[62235]: DEBUG nova.compute.manager [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing instance network info cache due to event network-changed-235e1a30-ec5f-4d9a-8085-8994ac4c8314. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.602791] env[62235]: DEBUG oslo_concurrency.lockutils [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.688646] env[62235]: DEBUG nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.696778] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.697377] env[62235]: DEBUG nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Instance network_info: |[{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.700172] env[62235]: DEBUG oslo_concurrency.lockutils [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.700722] env[62235]: DEBUG nova.network.neutron [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing network info cache for port 235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.702330] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:e1:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '235e1a30-ec5f-4d9a-8085-8994ac4c8314', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.711973] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Creating folder: Project (1e885dc2e5404662a081bb3855919b0d). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.720345] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2975c819-be79-4bd9-b361-e45fbe00f8a6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.730719] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.733794] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.734039] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.734263] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.734465] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.734658] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.734828] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.735112] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.735309] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.735523] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.735753] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.735944] env[62235]: DEBUG nova.virt.hardware [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.738848] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39c2efa-bc56-4eb1-ac3d-48d88b67de0a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.742052] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Created folder: Project (1e885dc2e5404662a081bb3855919b0d) in parent group-v273362. [ 811.742230] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Creating folder: Instances. Parent ref: group-v273416. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.743600] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.743943] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d74226be-0afd-4ce8-9dca-18d64ee2e843 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.753512] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3e926a-b8d5-4651-a548-50e396df4d8b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.759752] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Created folder: Instances in parent group-v273416. [ 811.760106] env[62235]: DEBUG oslo.service.loopingcall [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.760746] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.761040] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cba17867-4f3b-4817-a0fb-710b312be8d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.798860] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.798860] env[62235]: value = "task-1271610" [ 811.798860] env[62235]: _type = "Task" [ 811.798860] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.808179] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271610, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.973344] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.973877] env[62235]: DEBUG nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.976717] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.281s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.977953] env[62235]: INFO nova.compute.claims [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.015507] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271607, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453183} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.015808] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.016031] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.016517] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b968af1-9622-4da3-9a45-990151804859 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.022143] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 812.022143] env[62235]: value = "task-1271611" [ 812.022143] env[62235]: _type = "Task" [ 812.022143] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.029592] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.210780] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271606, 'name': ReconfigVM_Task, 'duration_secs': 0.624241} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.213095] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 590440de-4b3b-4019-b0c3-58b5ad329c88/590440de-4b3b-4019-b0c3-58b5ad329c88.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.213095] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b29122a-4d93-4326-990f-e081fe17527e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.223028] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 812.223028] env[62235]: value = "task-1271612" [ 812.223028] env[62235]: _type = "Task" [ 812.223028] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.236963] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271612, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.293946] env[62235]: DEBUG nova.compute.manager [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.296835] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309a63e4-6578-459b-bdaf-fecbb3f3421b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.318783] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271610, 'name': CreateVM_Task, 'duration_secs': 0.346626} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.319387] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.323023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.323023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.323023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.323023] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b64237c7-b7ea-4101-ba83-ac18ed8bc804 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.327983] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 812.327983] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52262c39-b5c4-027a-8a60-c0ccf79a2d6d" [ 812.327983] env[62235]: _type = "Task" [ 812.327983] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.348607] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52262c39-b5c4-027a-8a60-c0ccf79a2d6d, 'name': SearchDatastore_Task, 'duration_secs': 0.011524} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.348920] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.349171] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.349407] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.349555] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.349758] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.350036] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-996f2f4d-b295-426a-9c6e-de787e0df9d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.359797] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.360027] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.363215] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea77c733-e3c4-4311-885b-747dd7c22a50 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.368730] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 812.368730] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52098a8f-1b1e-5e16-5cef-bdeb8d14b3cb" [ 812.368730] env[62235]: _type = "Task" [ 812.368730] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.376499] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52098a8f-1b1e-5e16-5cef-bdeb8d14b3cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.424380] env[62235]: DEBUG nova.network.neutron [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Successfully updated port: aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.484340] env[62235]: DEBUG nova.compute.utils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.487102] env[62235]: DEBUG nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.487625] env[62235]: DEBUG nova.network.neutron [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.532994] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084137} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.533239] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.534224] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54ff5a7-0061-44b4-ab8a-a788c54bc499 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.566467] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.567038] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c964de5a-7124-46b8-960b-9f54788f1537 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.583440] env[62235]: DEBUG nova.policy [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '922955031b7f4ad0ba6daf2acc205018', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9bb60b466e948c5af70d80d3a1ad888', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.599038] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 812.599038] env[62235]: value = "task-1271613" [ 812.599038] env[62235]: _type = "Task" [ 812.599038] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.608561] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.609602] env[62235]: DEBUG nova.network.neutron [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updated VIF entry in instance network info cache for port 235e1a30-ec5f-4d9a-8085-8994ac4c8314. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.610106] env[62235]: DEBUG nova.network.neutron [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.734340] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271612, 'name': Rename_Task, 'duration_secs': 0.165722} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.736364] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.736364] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfd625e7-43b2-4488-85d2-22978f0bc376 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.744016] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 812.744016] env[62235]: value = "task-1271614" [ 812.744016] env[62235]: _type = "Task" [ 812.744016] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.751237] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271614, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.823022] env[62235]: INFO nova.compute.manager [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] instance snapshotting [ 812.825407] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9de8340-27f9-4f49-98d9-4335e66482ec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.844510] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf3df8f-9ff7-45f7-aad0-a32189279a78 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.879056] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52098a8f-1b1e-5e16-5cef-bdeb8d14b3cb, 'name': SearchDatastore_Task, 'duration_secs': 0.014346} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.880112] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48332ed3-750e-4e9c-b753-f203a907f449 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.885937] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 812.885937] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528bf96c-1791-f155-8515-79e703a0450f" [ 812.885937] env[62235]: _type = "Task" [ 812.885937] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.895343] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528bf96c-1791-f155-8515-79e703a0450f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.930826] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.931390] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.931542] env[62235]: DEBUG nova.network.neutron [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 812.990839] env[62235]: DEBUG nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.996252] env[62235]: DEBUG nova.network.neutron [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Successfully created port: 6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.110019] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271613, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.112689] env[62235]: DEBUG oslo_concurrency.lockutils [req-e5a53d6c-64cc-4a39-b054-4a256f32c239 req-37e759b2-7a94-4105-a83b-df2b9a916517 service nova] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.252534] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271614, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.254143] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4e9641-8c07-4257-81cb-8e74472e8c6a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.262152] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab52e11-440c-434f-bc9c-dda5d3725f69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.295207] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13f14f3-036a-4c05-aaab-1039159b1dc7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.303506] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3908eb-045f-45ae-b819-324338892f4a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.317348] env[62235]: DEBUG nova.compute.provider_tree [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.355601] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 813.355950] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7d62ca87-5240-4d6e-99a1-a8e098d4918c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.363185] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 813.363185] env[62235]: value = "task-1271615" [ 813.363185] env[62235]: _type = "Task" [ 813.363185] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.373604] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271615, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.396954] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528bf96c-1791-f155-8515-79e703a0450f, 'name': SearchDatastore_Task, 'duration_secs': 0.013395} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.397273] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.397558] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] ced2aa67-ce75-4af5-9c55-773835d05733/ced2aa67-ce75-4af5-9c55-773835d05733.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.397815] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0660645-9e13-4283-9d72-0d1f3174f094 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.404475] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 813.404475] env[62235]: value = "task-1271616" [ 813.404475] env[62235]: _type = "Task" [ 813.404475] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.411883] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.462356] env[62235]: DEBUG nova.network.neutron [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.611324] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271613, 'name': ReconfigVM_Task, 'duration_secs': 0.783193} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.611764] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.612715] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b2de539-daca-4de8-baca-1acbfc35d2e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.619590] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 813.619590] env[62235]: value = "task-1271617" [ 813.619590] env[62235]: _type = "Task" [ 813.619590] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.628617] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271617, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.642024] env[62235]: DEBUG nova.network.neutron [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updating instance_info_cache with network_info: [{"id": "aab67c7c-76aa-41fa-acba-76bec34f355c", "address": "fa:16:3e:88:57:a9", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab67c7c-76", "ovs_interfaceid": "aab67c7c-76aa-41fa-acba-76bec34f355c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.689940] env[62235]: DEBUG nova.compute.manager [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received event network-vif-plugged-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.690220] env[62235]: DEBUG oslo_concurrency.lockutils [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] Acquiring lock "d165f549-1759-41bf-86fd-b77793bd49c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.690378] env[62235]: DEBUG oslo_concurrency.lockutils [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] Lock "d165f549-1759-41bf-86fd-b77793bd49c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.690572] env[62235]: DEBUG oslo_concurrency.lockutils [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] Lock "d165f549-1759-41bf-86fd-b77793bd49c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.690996] env[62235]: DEBUG nova.compute.manager [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] No waiting events found dispatching network-vif-plugged-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 813.691349] env[62235]: WARNING nova.compute.manager [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received unexpected event network-vif-plugged-aab67c7c-76aa-41fa-acba-76bec34f355c for instance with vm_state building and task_state spawning. [ 813.691519] env[62235]: DEBUG nova.compute.manager [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.691726] env[62235]: DEBUG nova.compute.manager [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing instance network info cache due to event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 813.691854] env[62235]: DEBUG oslo_concurrency.lockutils [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] Acquiring lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.756348] env[62235]: DEBUG oslo_vmware.api [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271614, 'name': PowerOnVM_Task, 'duration_secs': 0.899042} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.756665] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.756887] env[62235]: INFO nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Took 9.22 seconds to spawn the instance on the hypervisor. [ 813.757086] env[62235]: DEBUG nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.758307] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300932fe-73d6-4001-93e4-6baba9759d61 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.820915] env[62235]: DEBUG nova.scheduler.client.report [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.873522] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271615, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.916125] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271616, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.000999] env[62235]: DEBUG nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.027654] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.028055] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.028226] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.028432] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.028589] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.028788] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.029044] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.029186] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.029374] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.029541] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.029756] env[62235]: DEBUG nova.virt.hardware [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.031043] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78641bda-fef1-4013-86c2-24ae830c4179 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.040059] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66287836-cc7d-47bc-b09f-dba3fe1d8fb4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.129254] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271617, 'name': Rename_Task, 'duration_secs': 0.416221} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.129553] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 814.129806] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3595e055-622b-4153-931a-6c2039e2acc6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.135564] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 814.135564] env[62235]: value = "task-1271618" [ 814.135564] env[62235]: _type = "Task" [ 814.135564] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.144060] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.144537] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Releasing lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.144858] env[62235]: DEBUG nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Instance network_info: |[{"id": "aab67c7c-76aa-41fa-acba-76bec34f355c", "address": "fa:16:3e:88:57:a9", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab67c7c-76", "ovs_interfaceid": "aab67c7c-76aa-41fa-acba-76bec34f355c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.145151] env[62235]: DEBUG oslo_concurrency.lockutils [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] Acquired lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.145335] env[62235]: DEBUG nova.network.neutron [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 814.146595] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:57:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '88651df2-0506-4f6c-b868-dd30a81f2b1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aab67c7c-76aa-41fa-acba-76bec34f355c', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.153871] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Creating folder: Project (4456e079eae64f41b1596821d41bd275). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.156709] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a93f6561-89eb-4a9e-86cb-3a5ec98da040 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.167603] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Created folder: Project (4456e079eae64f41b1596821d41bd275) in parent group-v273362. [ 814.167793] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Creating folder: Instances. Parent ref: group-v273420. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.167994] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28a4cc3d-0ef5-403c-b8cb-e4647636bd77 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.179456] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Created folder: Instances in parent group-v273420. [ 814.179725] env[62235]: DEBUG oslo.service.loopingcall [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.179917] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 814.180143] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7218072-4b87-4701-b389-fdb65c4479ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.201780] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.201780] env[62235]: value = "task-1271621" [ 814.201780] env[62235]: _type = "Task" [ 814.201780] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.209107] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271621, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.279604] env[62235]: INFO nova.compute.manager [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Took 41.12 seconds to build instance. [ 814.326668] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.327266] env[62235]: DEBUG nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.329907] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.332s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.333095] env[62235]: DEBUG nova.objects.instance [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lazy-loading 'resources' on Instance uuid d6324b5d-ec70-4bb4-b3c6-b1524abd0034 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.381423] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271615, 'name': CreateSnapshot_Task, 'duration_secs': 0.740912} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.381738] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 814.384233] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a8e0fa-f222-42e7-a138-75a6e169bdc5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.414211] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271616, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678055} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.414478] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] ced2aa67-ce75-4af5-9c55-773835d05733/ced2aa67-ce75-4af5-9c55-773835d05733.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.414690] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.414935] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-146b9a14-176b-401c-a98f-ba8475999cdf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.421388] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 814.421388] env[62235]: value = "task-1271622" [ 814.421388] env[62235]: _type = "Task" [ 814.421388] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.430168] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.651511] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271618, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.696322] env[62235]: DEBUG nova.network.neutron [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updated VIF entry in instance network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 814.696680] env[62235]: DEBUG nova.network.neutron [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updating instance_info_cache with network_info: [{"id": "aab67c7c-76aa-41fa-acba-76bec34f355c", "address": "fa:16:3e:88:57:a9", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab67c7c-76", "ovs_interfaceid": "aab67c7c-76aa-41fa-acba-76bec34f355c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.712120] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271621, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.782083] env[62235]: DEBUG oslo_concurrency.lockutils [None req-278ff0b0-edbc-402c-8d87-6bb57a507eb7 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.568s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.833020] env[62235]: DEBUG nova.compute.utils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.837917] env[62235]: DEBUG nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.838107] env[62235]: DEBUG nova.network.neutron [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 814.903204] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 814.906157] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1d7079fa-15df-4a71-89a1-f65c266869fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.914810] env[62235]: DEBUG nova.policy [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c1595adb0d14231bf83c33df8be1a8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e340ac03d3734399a92abe4ed014251c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 814.918186] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 814.918186] env[62235]: value = "task-1271623" [ 814.918186] env[62235]: _type = "Task" [ 814.918186] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.931330] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271623, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.936937] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096241} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.937275] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.938057] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad14a345-8485-48d0-8334-389aabe623f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.962778] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] ced2aa67-ce75-4af5-9c55-773835d05733/ced2aa67-ce75-4af5-9c55-773835d05733.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.963183] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59c264bf-fb6b-4df1-a186-641b720541b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.984776] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 814.984776] env[62235]: value = "task-1271624" [ 814.984776] env[62235]: _type = "Task" [ 814.984776] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.997305] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271624, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.997305] env[62235]: DEBUG nova.network.neutron [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Successfully updated port: 6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.149401] env[62235]: DEBUG oslo_vmware.api [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271618, 'name': PowerOnVM_Task, 'duration_secs': 0.861592} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.149679] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.149875] env[62235]: INFO nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Took 8.24 seconds to spawn the instance on the hypervisor. [ 815.150063] env[62235]: DEBUG nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.153124] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f301dc71-ce8e-4cb1-bcb1-59fec16c5298 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.185209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "590440de-4b3b-4019-b0c3-58b5ad329c88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.185445] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.185683] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "590440de-4b3b-4019-b0c3-58b5ad329c88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.185894] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.186158] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.189245] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68aae723-6a14-4b88-8ddf-e01aedcdad58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.192457] env[62235]: INFO nova.compute.manager [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Terminating instance [ 815.194612] env[62235]: DEBUG nova.compute.manager [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 815.194865] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.195600] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c708af-eb06-4c09-9d9c-65525caf7a8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.200509] env[62235]: DEBUG oslo_concurrency.lockutils [req-74a82ba4-aa20-4f5e-bd4e-d3e50d6c0117 req-d52b49bb-8f2e-4573-bd3d-c34c5e1a5f1c service nova] Releasing lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.202030] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd08868-b188-4374-9b8e-b4aa671a577a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.209014] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.210753] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d400292-c3ee-4873-945c-765d3d5b918f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.239439] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74875387-6519-4f7a-89d4-61657a005094 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.246922] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271621, 'name': CreateVM_Task, 'duration_secs': 0.581373} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.247329] env[62235]: DEBUG oslo_vmware.api [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 815.247329] env[62235]: value = "task-1271625" [ 815.247329] env[62235]: _type = "Task" [ 815.247329] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.247901] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.248560] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.249121] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.249190] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.256187] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2e64561-9115-41b5-8b1b-e21f7a70572b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.258934] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f212b0af-263e-4aab-b7ed-cd4b3eb8867b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.266479] env[62235]: DEBUG oslo_vmware.api [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.268360] env[62235]: DEBUG nova.network.neutron [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Successfully created port: 600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.280731] env[62235]: DEBUG nova.compute.provider_tree [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.285250] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 815.285250] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52786292-2f95-e48b-9a16-1185a14df7a0" [ 815.285250] env[62235]: _type = "Task" [ 815.285250] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.285957] env[62235]: DEBUG nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.295494] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52786292-2f95-e48b-9a16-1185a14df7a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.338709] env[62235]: DEBUG nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.428401] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271623, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.494792] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271624, 'name': ReconfigVM_Task, 'duration_secs': 0.351142} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.494792] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfigured VM instance instance-0000003f to attach disk [datastore2] ced2aa67-ce75-4af5-9c55-773835d05733/ced2aa67-ce75-4af5-9c55-773835d05733.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.495413] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24be6798-0a40-442e-9302-8db75c4c5787 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.502154] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "refresh_cache-b5a1da09-2137-4434-a1a4-174b196b61b0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.502303] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquired lock "refresh_cache-b5a1da09-2137-4434-a1a4-174b196b61b0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.502671] env[62235]: DEBUG nova.network.neutron [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.503879] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 815.503879] env[62235]: value = "task-1271626" [ 815.503879] env[62235]: _type = "Task" [ 815.503879] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.512397] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271626, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.670547] env[62235]: INFO nova.compute.manager [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Took 39.30 seconds to build instance. [ 815.714594] env[62235]: DEBUG nova.compute.manager [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Received event network-vif-plugged-6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.714795] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] Acquiring lock "b5a1da09-2137-4434-a1a4-174b196b61b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.714959] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.715284] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.715501] env[62235]: DEBUG nova.compute.manager [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] No waiting events found dispatching network-vif-plugged-6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 815.715672] env[62235]: WARNING nova.compute.manager [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Received unexpected event network-vif-plugged-6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 for instance with vm_state building and task_state spawning. [ 815.715868] env[62235]: DEBUG nova.compute.manager [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Received event network-changed-6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.716035] env[62235]: DEBUG nova.compute.manager [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Refreshing instance network info cache due to event network-changed-6a7bb767-b47e-4a64-ae3c-f05483ec7cd9. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 815.716210] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] Acquiring lock "refresh_cache-b5a1da09-2137-4434-a1a4-174b196b61b0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.758082] env[62235]: DEBUG oslo_vmware.api [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271625, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.783819] env[62235]: DEBUG nova.scheduler.client.report [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.798624] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52786292-2f95-e48b-9a16-1185a14df7a0, 'name': SearchDatastore_Task, 'duration_secs': 0.029866} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.798920] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.799159] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.799394] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.799551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.799712] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.800484] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02a0863f-0b2f-46d2-815d-d75dcfcb0f5d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.804231] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.808383] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.808568] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.809288] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de74b482-88f2-497c-a125-76172b7118c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.814538] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 815.814538] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5279bfd5-19bc-f5b2-60ff-3baa673aeaab" [ 815.814538] env[62235]: _type = "Task" [ 815.814538] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.822488] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5279bfd5-19bc-f5b2-60ff-3baa673aeaab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.928996] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271623, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.018785] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271626, 'name': Rename_Task, 'duration_secs': 0.13782} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.019132] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.019329] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f208f9c-d1ea-47c8-941e-5a2c6e0f25f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.027295] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 816.027295] env[62235]: value = "task-1271627" [ 816.027295] env[62235]: _type = "Task" [ 816.027295] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.036857] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271627, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.086344] env[62235]: DEBUG nova.network.neutron [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.173059] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4945f6b9-ee71-4959-95a4-9b1462923675 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.044s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.260154] env[62235]: DEBUG oslo_vmware.api [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271625, 'name': PowerOffVM_Task, 'duration_secs': 0.538558} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.260154] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.260154] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.260154] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a34754e7-543f-46e1-aa79-2f0ccdf2c649 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.290117] env[62235]: DEBUG nova.network.neutron [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Updating instance_info_cache with network_info: [{"id": "6a7bb767-b47e-4a64-ae3c-f05483ec7cd9", "address": "fa:16:3e:32:95:c3", "network": {"id": "8fbe2f1e-6d8e-483c-9f13-4780473e4240", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "52c5b3858c2a4a4c8151c52020553269", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a7bb767-b4", "ovs_interfaceid": "6a7bb767-b47e-4a64-ae3c-f05483ec7cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.291224] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.961s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.293732] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.963s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.295311] env[62235]: INFO nova.compute.claims [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.316459] env[62235]: INFO nova.scheduler.client.report [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Deleted allocations for instance d6324b5d-ec70-4bb4-b3c6-b1524abd0034 [ 816.330681] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5279bfd5-19bc-f5b2-60ff-3baa673aeaab, 'name': SearchDatastore_Task, 'duration_secs': 0.008114} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.332773] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ca8280a-18ec-4c95-a8ed-e3021c5e2fcf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.336192] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.336192] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.336328] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Deleting the datastore file [datastore2] 590440de-4b3b-4019-b0c3-58b5ad329c88 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.336692] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e4de4ce-925b-46ea-ab49-5cca4a8eeabf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.342588] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 816.342588] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52508359-ef95-7844-3564-78113bbcad55" [ 816.342588] env[62235]: _type = "Task" [ 816.342588] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.343965] env[62235]: DEBUG oslo_vmware.api [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for the task: (returnval){ [ 816.343965] env[62235]: value = "task-1271629" [ 816.343965] env[62235]: _type = "Task" [ 816.343965] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.348205] env[62235]: DEBUG nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.358977] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52508359-ef95-7844-3564-78113bbcad55, 'name': SearchDatastore_Task, 'duration_secs': 0.010097} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.362308] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.363166] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] d165f549-1759-41bf-86fd-b77793bd49c0/d165f549-1759-41bf-86fd-b77793bd49c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 816.363370] env[62235]: DEBUG oslo_vmware.api [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271629, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.363667] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf5bcdcc-d845-45a0-abce-b8e72946a2c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.371402] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 816.371402] env[62235]: value = "task-1271630" [ 816.371402] env[62235]: _type = "Task" [ 816.371402] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.380766] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.381027] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.381195] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.381400] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.381546] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.381693] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.381900] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.382070] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.382233] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.382390] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.382555] env[62235]: DEBUG nova.virt.hardware [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.383612] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9192cfe-1012-45bc-8929-12ed3f68ebf4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.389325] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.395527] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1387ebd0-1f68-4e5b-b387-fdcc57b3fbb4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.430614] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271623, 'name': CloneVM_Task} progress is 95%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.538619] env[62235]: DEBUG oslo_vmware.api [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271627, 'name': PowerOnVM_Task, 'duration_secs': 0.438631} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.538905] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.539147] env[62235]: INFO nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Took 7.25 seconds to spawn the instance on the hypervisor. [ 816.539426] env[62235]: DEBUG nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.540260] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ef820c-578d-4f3d-9662-8eb401f7eb67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.795046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Releasing lock "refresh_cache-b5a1da09-2137-4434-a1a4-174b196b61b0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.795615] env[62235]: DEBUG nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Instance network_info: |[{"id": "6a7bb767-b47e-4a64-ae3c-f05483ec7cd9", "address": "fa:16:3e:32:95:c3", "network": {"id": "8fbe2f1e-6d8e-483c-9f13-4780473e4240", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "52c5b3858c2a4a4c8151c52020553269", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a7bb767-b4", "ovs_interfaceid": "6a7bb767-b47e-4a64-ae3c-f05483ec7cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.796053] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] Acquired lock "refresh_cache-b5a1da09-2137-4434-a1a4-174b196b61b0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.796335] env[62235]: DEBUG nova.network.neutron [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Refreshing network info cache for port 6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.801027] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:95:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a7bb767-b47e-4a64-ae3c-f05483ec7cd9', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.811381] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Creating folder: Project (b9bb60b466e948c5af70d80d3a1ad888). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.821779] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2d0da3b-86c2-432f-9fc7-33b5a90620e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.829420] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98e29c0d-59f9-43dc-b689-a3a76c311468 tempest-ServerTagsTestJSON-1986622854 tempest-ServerTagsTestJSON-1986622854-project-member] Lock "d6324b5d-ec70-4bb4-b3c6-b1524abd0034" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.787s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.833661] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Created folder: Project (b9bb60b466e948c5af70d80d3a1ad888) in parent group-v273362. [ 816.834731] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Creating folder: Instances. Parent ref: group-v273424. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.834731] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8094442-c0f5-4f1b-b526-d81ccb4ea342 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.855775] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Created folder: Instances in parent group-v273424. [ 816.856069] env[62235]: DEBUG oslo.service.loopingcall [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.856716] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.856914] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6732b5e8-8b8b-43bd-b970-5db30cb8ddb6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.878944] env[62235]: DEBUG oslo_vmware.api [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Task: {'id': task-1271629, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170297} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.883290] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.883531] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.884772] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.884772] env[62235]: INFO nova.compute.manager [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Took 1.69 seconds to destroy the instance on the hypervisor. [ 816.884772] env[62235]: DEBUG oslo.service.loopingcall [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.886521] env[62235]: DEBUG nova.compute.manager [-] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.886683] env[62235]: DEBUG nova.network.neutron [-] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.888475] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.888475] env[62235]: value = "task-1271633" [ 816.888475] env[62235]: _type = "Task" [ 816.888475] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.896296] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271630, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.902010] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271633, 'name': CreateVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.929668] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271623, 'name': CloneVM_Task, 'duration_secs': 1.602691} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.929971] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Created linked-clone VM from snapshot [ 816.930780] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a34cf8-b06f-4229-9367-99e1261e68bd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.938751] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Uploading image 4b23c73d-e39b-4f1a-8817-627621a5ee39 {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 816.950829] env[62235]: DEBUG nova.network.neutron [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Successfully updated port: 600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.968922] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 816.968922] env[62235]: value = "vm-273423" [ 816.968922] env[62235]: _type = "VirtualMachine" [ 816.968922] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 816.969493] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-118e502b-a8f6-4642-bae1-877436a4047f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.975583] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lease: (returnval){ [ 816.975583] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525fc4f3-4008-eee0-83e7-a554c864b3d0" [ 816.975583] env[62235]: _type = "HttpNfcLease" [ 816.975583] env[62235]: } obtained for exporting VM: (result){ [ 816.975583] env[62235]: value = "vm-273423" [ 816.975583] env[62235]: _type = "VirtualMachine" [ 816.975583] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 816.975995] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the lease: (returnval){ [ 816.975995] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525fc4f3-4008-eee0-83e7-a554c864b3d0" [ 816.975995] env[62235]: _type = "HttpNfcLease" [ 816.975995] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 816.982728] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 816.982728] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525fc4f3-4008-eee0-83e7-a554c864b3d0" [ 816.982728] env[62235]: _type = "HttpNfcLease" [ 816.982728] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 817.059721] env[62235]: INFO nova.compute.manager [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Took 33.43 seconds to build instance. [ 817.173731] env[62235]: DEBUG nova.network.neutron [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Updated VIF entry in instance network info cache for port 6a7bb767-b47e-4a64-ae3c-f05483ec7cd9. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.174182] env[62235]: DEBUG nova.network.neutron [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Updating instance_info_cache with network_info: [{"id": "6a7bb767-b47e-4a64-ae3c-f05483ec7cd9", "address": "fa:16:3e:32:95:c3", "network": {"id": "8fbe2f1e-6d8e-483c-9f13-4780473e4240", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "52c5b3858c2a4a4c8151c52020553269", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a7bb767-b4", "ovs_interfaceid": "6a7bb767-b47e-4a64-ae3c-f05483ec7cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.395016] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271630, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537333} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.400306] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] d165f549-1759-41bf-86fd-b77793bd49c0/d165f549-1759-41bf-86fd-b77793bd49c0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 817.400306] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.407469] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-98fa3129-8a7d-495a-a9b4-59a3113ad7a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.418264] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271633, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.418264] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 817.418264] env[62235]: value = "task-1271635" [ 817.418264] env[62235]: _type = "Task" [ 817.418264] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.425522] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271635, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.454268] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.454268] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquired lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.454268] env[62235]: DEBUG nova.network.neutron [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.490898] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 817.490898] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525fc4f3-4008-eee0-83e7-a554c864b3d0" [ 817.490898] env[62235]: _type = "HttpNfcLease" [ 817.490898] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 817.491645] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 817.491645] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525fc4f3-4008-eee0-83e7-a554c864b3d0" [ 817.491645] env[62235]: _type = "HttpNfcLease" [ 817.491645] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 817.492469] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd302b64-2e09-4c33-b932-437324b35441 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.501249] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b0a59e-9e3f-2c57-7247-a9e0bd8cf9f2/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 817.501388] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b0a59e-9e3f-2c57-7247-a9e0bd8cf9f2/disk-0.vmdk for reading. {{(pid=62235) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 817.571914] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38a2097b-caf8-42bc-8a81-43bab81f43c9 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "ced2aa67-ce75-4af5-9c55-773835d05733" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.896s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.576587] env[62235]: DEBUG nova.compute.manager [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.576804] env[62235]: DEBUG nova.compute.manager [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing instance network info cache due to event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.576997] env[62235]: DEBUG oslo_concurrency.lockutils [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.577158] env[62235]: DEBUG oslo_concurrency.lockutils [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.577320] env[62235]: DEBUG nova.network.neutron [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.609572] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fddad650-4e76-40f9-9703-9203ee17f8b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.676960] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9ce6d56-dbe4-4b77-af08-d41f95906375 req-e1f437c2-51cf-4cfa-9e00-fed67d37d82b service nova] Releasing lock "refresh_cache-b5a1da09-2137-4434-a1a4-174b196b61b0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.705017] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8d4a62-1104-4db8-8707-e9dc89c4833c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.712784] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab818e0-482d-4688-9749-419c15f4ef33 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.744845] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699d2577-96c0-456b-a2ee-d432fc946623 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.748550] env[62235]: DEBUG nova.compute.manager [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received event network-vif-plugged-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.748757] env[62235]: DEBUG oslo_concurrency.lockutils [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] Acquiring lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.748959] env[62235]: DEBUG oslo_concurrency.lockutils [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.749281] env[62235]: DEBUG oslo_concurrency.lockutils [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.749361] env[62235]: DEBUG nova.compute.manager [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] No waiting events found dispatching network-vif-plugged-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.749538] env[62235]: WARNING nova.compute.manager [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received unexpected event network-vif-plugged-600609cd-3b8f-4683-9136-308a989872ec for instance with vm_state building and task_state spawning. [ 817.749617] env[62235]: DEBUG nova.compute.manager [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received event network-changed-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.749751] env[62235]: DEBUG nova.compute.manager [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing instance network info cache due to event network-changed-600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.749913] env[62235]: DEBUG oslo_concurrency.lockutils [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] Acquiring lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.756080] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cfc0927-4a83-45d1-9f24-566cbbe27315 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.770184] env[62235]: DEBUG nova.compute.provider_tree [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.907742] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271633, 'name': CreateVM_Task, 'duration_secs': 0.585093} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.907941] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.908712] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.909133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.912073] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.912073] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9c7a653-4dc7-4623-85c6-2d8c24a137bb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.915010] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 817.915010] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52af11ca-18aa-9608-a1b6-f3c5ac617c87" [ 817.915010] env[62235]: _type = "Task" [ 817.915010] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.927783] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271635, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078413} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.930513] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.931059] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52af11ca-18aa-9608-a1b6-f3c5ac617c87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.931683] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1440dbe7-5f11-4801-b168-14d6c0e5043d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.962970] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] d165f549-1759-41bf-86fd-b77793bd49c0/d165f549-1759-41bf-86fd-b77793bd49c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.965352] env[62235]: DEBUG nova.network.neutron [-] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.968430] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56e31296-9a9e-4dfb-ad4c-f0ae61327587 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.992287] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 817.992287] env[62235]: value = "task-1271636" [ 817.992287] env[62235]: _type = "Task" [ 817.992287] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.000969] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271636, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.043422] env[62235]: DEBUG nova.network.neutron [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.248047] env[62235]: DEBUG nova.network.neutron [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [{"id": "600609cd-3b8f-4683-9136-308a989872ec", "address": "fa:16:3e:9f:b2:2e", "network": {"id": "e3892181-6713-45be-9572-f39a5a3955b2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-769806034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e340ac03d3734399a92abe4ed014251c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600609cd-3b", "ovs_interfaceid": "600609cd-3b8f-4683-9136-308a989872ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.274162] env[62235]: DEBUG nova.scheduler.client.report [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.429234] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52af11ca-18aa-9608-a1b6-f3c5ac617c87, 'name': SearchDatastore_Task, 'duration_secs': 0.012451} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.429578] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.429790] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.430038] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.430194] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.430414] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.430964] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16c213ce-8e5e-492e-bd2d-d3532e1628fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.440209] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.440399] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.441148] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b755206-af69-46a4-a396-7f4a642d4d4e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.446657] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 818.446657] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a075d7-ba01-8ac4-0547-fdaa54fa67c0" [ 818.446657] env[62235]: _type = "Task" [ 818.446657] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.455186] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a075d7-ba01-8ac4-0547-fdaa54fa67c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.487158] env[62235]: INFO nova.compute.manager [-] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Took 1.60 seconds to deallocate network for instance. [ 818.507674] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271636, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.752501] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Releasing lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.753154] env[62235]: DEBUG nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Instance network_info: |[{"id": "600609cd-3b8f-4683-9136-308a989872ec", "address": "fa:16:3e:9f:b2:2e", "network": {"id": "e3892181-6713-45be-9572-f39a5a3955b2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-769806034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e340ac03d3734399a92abe4ed014251c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600609cd-3b", "ovs_interfaceid": "600609cd-3b8f-4683-9136-308a989872ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.753523] env[62235]: DEBUG oslo_concurrency.lockutils [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] Acquired lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.753716] env[62235]: DEBUG nova.network.neutron [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing network info cache for port 600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.759021] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:b2:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca16b6db-4f74-424c-9d36-925ad82cbdd2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '600609cd-3b8f-4683-9136-308a989872ec', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.767894] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Creating folder: Project (e340ac03d3734399a92abe4ed014251c). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.769238] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff502d93-025c-45ad-a4f1-56c1ae6cfd66 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.781538] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.782136] env[62235]: DEBUG nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.785438] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Created folder: Project (e340ac03d3734399a92abe4ed014251c) in parent group-v273362. [ 818.787262] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Creating folder: Instances. Parent ref: group-v273427. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.788009] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.245s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.788383] env[62235]: DEBUG nova.objects.instance [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'resources' on Instance uuid afa35ae0-40c6-4eaa-80f5-c1ada27f1871 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.793018] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ca976b2-06fa-4605-b0ee-8cee2c946f05 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.803163] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Created folder: Instances in parent group-v273427. [ 818.803163] env[62235]: DEBUG oslo.service.loopingcall [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.803338] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.803540] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3afe85e-726e-4d4d-9dc9-0c1cf6bcc0a9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.826445] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.826445] env[62235]: value = "task-1271639" [ 818.826445] env[62235]: _type = "Task" [ 818.826445] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.838744] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271639, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.914438] env[62235]: DEBUG nova.network.neutron [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updated VIF entry in instance network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 818.915918] env[62235]: DEBUG nova.network.neutron [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.961043] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a075d7-ba01-8ac4-0547-fdaa54fa67c0, 'name': SearchDatastore_Task, 'duration_secs': 0.02178} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.961556] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aed71311-d7b7-453b-8f1b-727010dd878f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.967264] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 818.967264] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528b8e8f-b378-2fdb-b88b-08785af7b6ec" [ 818.967264] env[62235]: _type = "Task" [ 818.967264] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.975399] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528b8e8f-b378-2fdb-b88b-08785af7b6ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.999457] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.009029] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271636, 'name': ReconfigVM_Task, 'duration_secs': 0.770336} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.009029] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Reconfigured VM instance instance-0000003e to attach disk [datastore2] d165f549-1759-41bf-86fd-b77793bd49c0/d165f549-1759-41bf-86fd-b77793bd49c0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.009363] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3863aa62-20e9-46c7-aba1-46d037591938 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.016197] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 819.016197] env[62235]: value = "task-1271640" [ 819.016197] env[62235]: _type = "Task" [ 819.016197] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.025687] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271640, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.289712] env[62235]: DEBUG nova.compute.utils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.291277] env[62235]: DEBUG nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.291460] env[62235]: DEBUG nova.network.neutron [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.342193] env[62235]: DEBUG nova.policy [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eff208ffe4814720bb6cc29dd1cdb6c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1f2551ad3cfb4a0d8112f85d4e95fe64', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.350168] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271639, 'name': CreateVM_Task, 'duration_secs': 0.359196} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.350168] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.350710] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.350888] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.351242] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.351518] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bede7cd-a194-4632-91b4-06b91dcc4dbc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.359333] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 819.359333] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52077374-ac3a-b9de-eef4-8bf8c78c6679" [ 819.359333] env[62235]: _type = "Task" [ 819.359333] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.371171] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52077374-ac3a-b9de-eef4-8bf8c78c6679, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.418731] env[62235]: DEBUG oslo_concurrency.lockutils [req-dd1455cf-7985-4bfa-b589-a26059566fca req-8c341d29-94e5-449c-b827-4b5031cf1e1a service nova] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.482820] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528b8e8f-b378-2fdb-b88b-08785af7b6ec, 'name': SearchDatastore_Task, 'duration_secs': 0.012012} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.483124] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.483412] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b5a1da09-2137-4434-a1a4-174b196b61b0/b5a1da09-2137-4434-a1a4-174b196b61b0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.483667] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65664845-23b9-4938-b12f-314c312c910b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.493717] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 819.493717] env[62235]: value = "task-1271641" [ 819.493717] env[62235]: _type = "Task" [ 819.493717] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.503652] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.534269] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271640, 'name': Rename_Task, 'duration_secs': 0.183327} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.539462] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.540217] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09bc7f90-7bd9-47e3-a929-918a4de9390b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.548492] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 819.548492] env[62235]: value = "task-1271642" [ 819.548492] env[62235]: _type = "Task" [ 819.548492] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.558468] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271642, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.563935] env[62235]: DEBUG nova.compute.manager [req-814b0b49-5241-4a3e-b968-f987e8325401 req-96d3a475-f717-4aff-9620-8d30a472ed16 service nova] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Received event network-vif-deleted-fa915021-a55c-4801-90c1-4c80e3d88c70 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.596896] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4774ea14-6930-48dc-989f-4fc6c600db3d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.605058] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837b65e5-f1e7-48dc-b6b1-6db31c8aa98d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.639951] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca0e3ba-ad31-4834-988c-e81232a7854e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.648776] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46aa31d3-8976-4a51-b4ce-28bb5242fd8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.663177] env[62235]: DEBUG nova.compute.provider_tree [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.773177] env[62235]: DEBUG nova.network.neutron [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Successfully created port: 0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.789243] env[62235]: DEBUG nova.network.neutron [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updated VIF entry in instance network info cache for port 600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.789641] env[62235]: DEBUG nova.network.neutron [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [{"id": "600609cd-3b8f-4683-9136-308a989872ec", "address": "fa:16:3e:9f:b2:2e", "network": {"id": "e3892181-6713-45be-9572-f39a5a3955b2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-769806034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e340ac03d3734399a92abe4ed014251c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600609cd-3b", "ovs_interfaceid": "600609cd-3b8f-4683-9136-308a989872ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.795447] env[62235]: DEBUG nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.872052] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52077374-ac3a-b9de-eef4-8bf8c78c6679, 'name': SearchDatastore_Task, 'duration_secs': 0.01607} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.872052] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.872052] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.872052] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.872255] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.872255] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.872313] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7c84a5b-1fb6-482a-95f5-3e14ca78c36b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.882293] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.882497] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.883691] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9354bcff-745d-4370-bc44-561f76d93caa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.891889] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 819.891889] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b8783e-16df-722c-9114-3a1f6c4b88de" [ 819.891889] env[62235]: _type = "Task" [ 819.891889] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.900702] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b8783e-16df-722c-9114-3a1f6c4b88de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.939104] env[62235]: DEBUG nova.compute.manager [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-changed-235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.939344] env[62235]: DEBUG nova.compute.manager [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing instance network info cache due to event network-changed-235e1a30-ec5f-4d9a-8085-8994ac4c8314. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.939654] env[62235]: DEBUG oslo_concurrency.lockutils [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.939741] env[62235]: DEBUG oslo_concurrency.lockutils [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.939891] env[62235]: DEBUG nova.network.neutron [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing network info cache for port 235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.006023] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271641, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.063062] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271642, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.169398] env[62235]: DEBUG nova.scheduler.client.report [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.293759] env[62235]: DEBUG oslo_concurrency.lockutils [req-814ca57d-3e4f-49f5-9483-c3f7769a3c0c req-e2a489f6-88f4-464a-a965-87e0d48f08a6 service nova] Releasing lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.404459] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b8783e-16df-722c-9114-3a1f6c4b88de, 'name': SearchDatastore_Task, 'duration_secs': 0.034508} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.405388] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf88b3b0-96e9-4b9f-9602-4793ce4db623 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.411496] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 820.411496] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dd6cfc-79e6-a1b5-2bce-290767574ddb" [ 820.411496] env[62235]: _type = "Task" [ 820.411496] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.420507] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dd6cfc-79e6-a1b5-2bce-290767574ddb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.508464] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.745434} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.508886] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b5a1da09-2137-4434-a1a4-174b196b61b0/b5a1da09-2137-4434-a1a4-174b196b61b0.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.509303] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.509601] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01371a3f-328b-4123-9574-211b3bffe311 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.517107] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 820.517107] env[62235]: value = "task-1271643" [ 820.517107] env[62235]: _type = "Task" [ 820.517107] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.526712] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.559031] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271642, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.675322] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.678687] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.380s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.678687] env[62235]: DEBUG nova.objects.instance [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lazy-loading 'resources' on Instance uuid dd7268d1-1da5-4e5f-9d45-816df147b564 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.710318] env[62235]: INFO nova.scheduler.client.report [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted allocations for instance afa35ae0-40c6-4eaa-80f5-c1ada27f1871 [ 820.735325] env[62235]: DEBUG nova.network.neutron [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updated VIF entry in instance network info cache for port 235e1a30-ec5f-4d9a-8085-8994ac4c8314. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.735714] env[62235]: DEBUG nova.network.neutron [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.808154] env[62235]: DEBUG nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.834209] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.834575] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.834759] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.834963] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.835146] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.835307] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.835545] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.835726] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.835934] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.836124] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.836309] env[62235]: DEBUG nova.virt.hardware [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.837241] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b841422-e893-4182-a514-526a20b08fb9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.845965] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbf05de-61ca-422e-afa3-79d5c4cb1d5a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.924429] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dd6cfc-79e6-a1b5-2bce-290767574ddb, 'name': SearchDatastore_Task, 'duration_secs': 0.02474} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.924872] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.925644] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/876b7602-ccf0-4aed-90e2-d07fbf548fb6.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.926053] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42afee25-914d-4ef0-b1ea-5467cf96633f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.937655] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 820.937655] env[62235]: value = "task-1271644" [ 820.937655] env[62235]: _type = "Task" [ 820.937655] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.952434] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.027795] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.165763} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.028647] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.029092] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b47831-c253-4d10-bf0d-a67463846bcc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.054619] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] b5a1da09-2137-4434-a1a4-174b196b61b0/b5a1da09-2137-4434-a1a4-174b196b61b0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.055053] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d0e6295-9947-45ae-9113-14facbdbb4d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.081579] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271642, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.083384] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 821.083384] env[62235]: value = "task-1271645" [ 821.083384] env[62235]: _type = "Task" [ 821.083384] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.091687] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.221527] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47eef002-6df4-458c-a68a-8070076956bb tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "afa35ae0-40c6-4eaa-80f5-c1ada27f1871" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.747s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.239272] env[62235]: DEBUG oslo_concurrency.lockutils [req-7f2eb25f-eaed-4a09-934d-4b5e04648339 req-17463022-32f5-45fd-b486-64a03d3019ac service nova] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.451896] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271644, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.480078] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96d723f-2e3f-4168-aa40-663ff74c6576 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.489239] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8512f411-4a69-4273-80a3-4071ee795eee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.530084] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdc83dc-ab29-40f6-b970-5cf095b4148c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.536187] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe53a783-b303-4b5a-b81e-d54508c36c93 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.560193] env[62235]: DEBUG nova.compute.provider_tree [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.576322] env[62235]: DEBUG oslo_vmware.api [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271642, 'name': PowerOnVM_Task, 'duration_secs': 1.561196} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.576730] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.576878] env[62235]: INFO nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Took 9.89 seconds to spawn the instance on the hypervisor. [ 821.577113] env[62235]: DEBUG nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.577985] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fa07b4-b75c-4fd2-a78d-6aa20f2ef96f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.599973] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271645, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.783648] env[62235]: DEBUG nova.compute.manager [req-12e3b91b-fa91-4c5a-add6-8ff468b8aa87 req-15e4b8ea-1e43-44b6-93e3-cc60f2e72f07 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Received event network-vif-plugged-0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.783648] env[62235]: DEBUG oslo_concurrency.lockutils [req-12e3b91b-fa91-4c5a-add6-8ff468b8aa87 req-15e4b8ea-1e43-44b6-93e3-cc60f2e72f07 service nova] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.783648] env[62235]: DEBUG oslo_concurrency.lockutils [req-12e3b91b-fa91-4c5a-add6-8ff468b8aa87 req-15e4b8ea-1e43-44b6-93e3-cc60f2e72f07 service nova] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.783987] env[62235]: DEBUG oslo_concurrency.lockutils [req-12e3b91b-fa91-4c5a-add6-8ff468b8aa87 req-15e4b8ea-1e43-44b6-93e3-cc60f2e72f07 service nova] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.784118] env[62235]: DEBUG nova.compute.manager [req-12e3b91b-fa91-4c5a-add6-8ff468b8aa87 req-15e4b8ea-1e43-44b6-93e3-cc60f2e72f07 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] No waiting events found dispatching network-vif-plugged-0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 821.784220] env[62235]: WARNING nova.compute.manager [req-12e3b91b-fa91-4c5a-add6-8ff468b8aa87 req-15e4b8ea-1e43-44b6-93e3-cc60f2e72f07 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Received unexpected event network-vif-plugged-0d781479-5511-42b0-9aa5-4efbe57d8d58 for instance with vm_state building and task_state spawning. [ 821.846882] env[62235]: DEBUG nova.network.neutron [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Successfully updated port: 0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.948424] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.82431} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.948715] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/876b7602-ccf0-4aed-90e2-d07fbf548fb6.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.948941] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.949233] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08d3eca3-79f1-4c9a-b3b2-b2880d373ad3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.955783] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 821.955783] env[62235]: value = "task-1271646" [ 821.955783] env[62235]: _type = "Task" [ 821.955783] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.965138] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.072133] env[62235]: DEBUG nova.scheduler.client.report [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.100862] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271645, 'name': ReconfigVM_Task, 'duration_secs': 0.798535} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.103048] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Reconfigured VM instance instance-00000040 to attach disk [datastore1] b5a1da09-2137-4434-a1a4-174b196b61b0/b5a1da09-2137-4434-a1a4-174b196b61b0.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.104028] env[62235]: INFO nova.compute.manager [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Took 38.48 seconds to build instance. [ 822.104940] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64dd8ea8-09ec-4a57-bdc2-9b04146343dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.113181] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 822.113181] env[62235]: value = "task-1271647" [ 822.113181] env[62235]: _type = "Task" [ 822.113181] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.122710] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271647, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.350469] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.350639] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquired lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.351292] env[62235]: DEBUG nova.network.neutron [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.450761] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "9474de53-1964-4a52-8783-2bac5773b6c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.451108] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.467662] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128238} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.468671] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.469620] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2934772b-141f-4c27-a6f8-bcf8a2141d25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.499051] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/876b7602-ccf0-4aed-90e2-d07fbf548fb6.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.499509] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a1338d8-96c1-4da7-92ac-3221e9f964cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.521877] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 822.521877] env[62235]: value = "task-1271648" [ 822.521877] env[62235]: _type = "Task" [ 822.521877] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.531867] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271648, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.576756] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.899s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.583977] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.621s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.587822] env[62235]: INFO nova.compute.claims [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.609010] env[62235]: DEBUG oslo_concurrency.lockutils [None req-baf4c85d-8d10-423b-8229-ca8dd94a7449 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "d165f549-1759-41bf-86fd-b77793bd49c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.579s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.616283] env[62235]: INFO nova.scheduler.client.report [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Deleted allocations for instance dd7268d1-1da5-4e5f-9d45-816df147b564 [ 822.634820] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271647, 'name': Rename_Task, 'duration_secs': 0.180186} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.634820] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 822.634820] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5e8efd3-b0f8-450d-81d5-e411ab8febf9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.640144] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 822.640144] env[62235]: value = "task-1271649" [ 822.640144] env[62235]: _type = "Task" [ 822.640144] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.652247] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.957876] env[62235]: DEBUG nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.007189] env[62235]: DEBUG nova.network.neutron [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.037617] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271648, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.126854] env[62235]: DEBUG oslo_concurrency.lockutils [None req-813518ed-ccf9-4b09-824c-8398b241ef81 tempest-MultipleCreateTestJSON-1824292726 tempest-MultipleCreateTestJSON-1824292726-project-member] Lock "dd7268d1-1da5-4e5f-9d45-816df147b564" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.714s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.155725] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271649, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.305792] env[62235]: DEBUG nova.network.neutron [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updating instance_info_cache with network_info: [{"id": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "address": "fa:16:3e:03:4a:67", "network": {"id": "df25d9d9-5721-4c36-b75a-2cf35c302b03", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-520889908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f2551ad3cfb4a0d8112f85d4e95fe64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d781479-55", "ovs_interfaceid": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.505009] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.535200] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271648, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.658284] env[62235]: DEBUG oslo_vmware.api [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271649, 'name': PowerOnVM_Task, 'duration_secs': 0.57636} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.658284] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 823.658284] env[62235]: INFO nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Took 9.66 seconds to spawn the instance on the hypervisor. [ 823.659242] env[62235]: DEBUG nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.663117] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5084bf9d-6fac-4a5a-bf80-95d3b012614b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.810216] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Releasing lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.810560] env[62235]: DEBUG nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Instance network_info: |[{"id": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "address": "fa:16:3e:03:4a:67", "network": {"id": "df25d9d9-5721-4c36-b75a-2cf35c302b03", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-520889908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f2551ad3cfb4a0d8112f85d4e95fe64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d781479-55", "ovs_interfaceid": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.811030] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:4a:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b67e519-46cf-44ce-b670-4ba4c0c5b658', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0d781479-5511-42b0-9aa5-4efbe57d8d58', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.822910] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Creating folder: Project (1f2551ad3cfb4a0d8112f85d4e95fe64). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.825346] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-136d0a58-142b-4d94-aba9-c8edfe91221b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.830818] env[62235]: DEBUG nova.compute.manager [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Received event network-changed-0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.831089] env[62235]: DEBUG nova.compute.manager [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Refreshing instance network info cache due to event network-changed-0d781479-5511-42b0-9aa5-4efbe57d8d58. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.831344] env[62235]: DEBUG oslo_concurrency.lockutils [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] Acquiring lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.831493] env[62235]: DEBUG oslo_concurrency.lockutils [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] Acquired lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.831657] env[62235]: DEBUG nova.network.neutron [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Refreshing network info cache for port 0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.837030] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Created folder: Project (1f2551ad3cfb4a0d8112f85d4e95fe64) in parent group-v273362. [ 823.837448] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Creating folder: Instances. Parent ref: group-v273430. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 823.837616] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a65a7de-d6eb-4418-8e75-6b9950a6293f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.852271] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Created folder: Instances in parent group-v273430. [ 823.852562] env[62235]: DEBUG oslo.service.loopingcall [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.852969] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 823.855619] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57a54c47-baa5-424c-9bf7-7dbcbcd16843 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.883481] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.883481] env[62235]: value = "task-1271652" [ 823.883481] env[62235]: _type = "Task" [ 823.883481] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.892149] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271652, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.920260] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f19256-9a44-4ebe-bea6-752fdec78b23 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.929628] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f834bfe0-2aa4-452c-a01b-9fa3251e24cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.979636] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de5dcdb-863c-47f1-8598-ef6a5ebc95f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.993247] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf44f71-fe09-4dff-8eb8-3b582205e1e0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.008376] env[62235]: DEBUG nova.compute.provider_tree [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.035252] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271648, 'name': ReconfigVM_Task, 'duration_secs': 1.455321} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.035668] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/876b7602-ccf0-4aed-90e2-d07fbf548fb6.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.036523] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9605b7d-12da-4ef6-aee2-16eabb928457 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.047810] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 824.047810] env[62235]: value = "task-1271653" [ 824.047810] env[62235]: _type = "Task" [ 824.047810] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.057186] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271653, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.185759] env[62235]: INFO nova.compute.manager [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Took 39.44 seconds to build instance. [ 824.397218] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271652, 'name': CreateVM_Task, 'duration_secs': 0.422478} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.400694] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.400694] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.400694] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.400694] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.400694] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd9bcb02-9d08-4dd3-bb40-9a6701a1990d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.405889] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 824.405889] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a12bcc-422b-3870-dc07-fe42d738cb26" [ 824.405889] env[62235]: _type = "Task" [ 824.405889] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.415775] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a12bcc-422b-3870-dc07-fe42d738cb26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.511644] env[62235]: DEBUG nova.scheduler.client.report [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.554857] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271653, 'name': Rename_Task, 'duration_secs': 0.204223} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.555221] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 824.555488] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9823553e-8a59-4f45-9d41-fa5d7e901ec7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.562241] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 824.562241] env[62235]: value = "task-1271654" [ 824.562241] env[62235]: _type = "Task" [ 824.562241] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.571546] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271654, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.689139] env[62235]: DEBUG oslo_concurrency.lockutils [None req-711b0691-ab23-4ea1-a075-463b877473cd tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.193s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.817537] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "b7801744-5552-49b8-a95d-41b67bdf2380" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.817537] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.834479] env[62235]: DEBUG nova.network.neutron [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updated VIF entry in instance network info cache for port 0d781479-5511-42b0-9aa5-4efbe57d8d58. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.834793] env[62235]: DEBUG nova.network.neutron [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updating instance_info_cache with network_info: [{"id": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "address": "fa:16:3e:03:4a:67", "network": {"id": "df25d9d9-5721-4c36-b75a-2cf35c302b03", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-520889908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f2551ad3cfb4a0d8112f85d4e95fe64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d781479-55", "ovs_interfaceid": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.927974] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a12bcc-422b-3870-dc07-fe42d738cb26, 'name': SearchDatastore_Task, 'duration_secs': 0.015892} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.927974] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.928400] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.928995] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.929190] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.929396] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.929705] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ba4ec6e-3094-4c0f-ad5e-d006a1536964 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.940597] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.940856] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 824.942018] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc2b50ac-ffe9-4913-9e6c-861f85f76863 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.948661] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 824.948661] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525b34df-58f0-34b5-4ceb-d0a6452fc299" [ 824.948661] env[62235]: _type = "Task" [ 824.948661] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.959259] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525b34df-58f0-34b5-4ceb-d0a6452fc299, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.018474] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.019102] env[62235]: DEBUG nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.022227] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.643s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.022436] env[62235]: DEBUG nova.objects.instance [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 825.074624] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271654, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.320488] env[62235]: DEBUG nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.338166] env[62235]: DEBUG oslo_concurrency.lockutils [req-5e95c199-a2e1-4972-a5cc-8258b5ed2d6b req-30d4848c-e832-4e28-aac7-88f7fbc9d9a3 service nova] Releasing lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.422539] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "b5a1da09-2137-4434-a1a4-174b196b61b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.422832] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.423057] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "b5a1da09-2137-4434-a1a4-174b196b61b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.423267] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.423768] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.426949] env[62235]: INFO nova.compute.manager [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Terminating instance [ 825.428899] env[62235]: DEBUG nova.compute.manager [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.429157] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 825.430013] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7e7d5e-631b-4351-8d5d-5742519d87dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.438394] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 825.438670] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcc4aca5-98c5-4a22-9c39-6328b6764216 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.444846] env[62235]: DEBUG oslo_vmware.api [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 825.444846] env[62235]: value = "task-1271655" [ 825.444846] env[62235]: _type = "Task" [ 825.444846] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.456158] env[62235]: DEBUG oslo_vmware.api [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271655, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.462103] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525b34df-58f0-34b5-4ceb-d0a6452fc299, 'name': SearchDatastore_Task, 'duration_secs': 0.012289} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.465281] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a31989e8-527b-4e68-8841-4f2e8f24f821 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.469273] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 825.469273] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52420e03-272e-51f0-fe3a-9ebed86fbb32" [ 825.469273] env[62235]: _type = "Task" [ 825.469273] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.478129] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52420e03-272e-51f0-fe3a-9ebed86fbb32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.530436] env[62235]: DEBUG nova.compute.utils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.536186] env[62235]: DEBUG nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.537473] env[62235]: DEBUG nova.network.neutron [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 825.577242] env[62235]: DEBUG oslo_vmware.api [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271654, 'name': PowerOnVM_Task, 'duration_secs': 0.536344} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.577943] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.577943] env[62235]: INFO nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Took 9.23 seconds to spawn the instance on the hypervisor. [ 825.578267] env[62235]: DEBUG nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.579547] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a883a75-eb23-4c80-8a72-e3bf061d029b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.598498] env[62235]: DEBUG nova.policy [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a58ee2911afa4a6fabd2d6eabd9e6b8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '940941e982494a4caeebda900b3e0b8c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.854333] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.954889] env[62235]: DEBUG oslo_vmware.api [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271655, 'name': PowerOffVM_Task, 'duration_secs': 0.23365} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.955246] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 825.955491] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 825.955694] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-feb2a0dc-afc0-4db9-abab-9bb3ed2b8216 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.979354] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52420e03-272e-51f0-fe3a-9ebed86fbb32, 'name': SearchDatastore_Task, 'duration_secs': 0.019621} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.980213] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.980213] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 06ec6b9c-bef0-4f96-a185-4315961de7f7/06ec6b9c-bef0-4f96-a185-4315961de7f7.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.980398] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35515e9d-f916-498e-8587-e33d4708e3a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.987629] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 825.987629] env[62235]: value = "task-1271657" [ 825.987629] env[62235]: _type = "Task" [ 825.987629] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.999195] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271657, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.016372] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 826.016638] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 826.016884] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Deleting the datastore file [datastore1] b5a1da09-2137-4434-a1a4-174b196b61b0 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.017193] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2febcc02-e19a-4d80-ba6a-831c859c02f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.024509] env[62235]: DEBUG oslo_vmware.api [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for the task: (returnval){ [ 826.024509] env[62235]: value = "task-1271658" [ 826.024509] env[62235]: _type = "Task" [ 826.024509] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.040050] env[62235]: DEBUG nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.042918] env[62235]: DEBUG oslo_vmware.api [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.044227] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ee8af272-d953-4882-bc3b-d2283f54b756 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.045245] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.409s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.045510] env[62235]: DEBUG nova.objects.instance [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lazy-loading 'resources' on Instance uuid b4a3777f-2ee5-464b-84eb-0722d59f2b44 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 826.063226] env[62235]: DEBUG nova.network.neutron [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Successfully created port: 5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.114147] env[62235]: INFO nova.compute.manager [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Took 40.44 seconds to build instance. [ 826.500302] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271657, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.535803] env[62235]: DEBUG oslo_vmware.api [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Task: {'id': task-1271658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301263} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.536264] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.536659] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 826.536776] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 826.537035] env[62235]: INFO nova.compute.manager [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 826.537433] env[62235]: DEBUG oslo.service.loopingcall [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.537692] env[62235]: DEBUG nova.compute.manager [-] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.537893] env[62235]: DEBUG nova.network.neutron [-] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 826.616908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-257a2f5a-c3d7-47ae-a09c-e5008d809bed tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.934s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.678308] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b0a59e-9e3f-2c57-7247-a9e0bd8cf9f2/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 826.679369] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684fa527-85e7-4ae7-a47a-2e7fcd1520c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.690986] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b0a59e-9e3f-2c57-7247-a9e0bd8cf9f2/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 826.691615] env[62235]: ERROR oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b0a59e-9e3f-2c57-7247-a9e0bd8cf9f2/disk-0.vmdk due to incomplete transfer. [ 826.692164] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d5370fba-29cc-46f3-90b3-bb088158f2b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.700154] env[62235]: DEBUG oslo_vmware.rw_handles [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b0a59e-9e3f-2c57-7247-a9e0bd8cf9f2/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 826.700402] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Uploaded image 4b23c73d-e39b-4f1a-8817-627621a5ee39 to the Glance image server {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 826.704011] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 826.704435] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c908ea59-2c20-4a73-a658-12974802c8f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.714137] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 826.714137] env[62235]: value = "task-1271659" [ 826.714137] env[62235]: _type = "Task" [ 826.714137] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.722648] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271659, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.842431] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1917b1-d764-405e-9138-d866b97b9cef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.850564] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449f5409-6f25-40d5-8c30-025efb835f4a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.889455] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a135e9-046d-4ac4-bd48-6a2833a7800e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.903717] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8e177b-5ed6-4e4e-a9ef-2962e7ef5d30 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.920553] env[62235]: DEBUG nova.compute.provider_tree [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.007407] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542386} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.007677] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 06ec6b9c-bef0-4f96-a185-4315961de7f7/06ec6b9c-bef0-4f96-a185-4315961de7f7.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.007888] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.008440] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31f156ba-d6b8-4587-9302-22e24afe5e98 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.015809] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 827.015809] env[62235]: value = "task-1271660" [ 827.015809] env[62235]: _type = "Task" [ 827.015809] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.023792] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.051311] env[62235]: DEBUG nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.081777] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.082047] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.082220] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.082408] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.082561] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.082714] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.082925] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.083101] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.083278] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.083453] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.083619] env[62235]: DEBUG nova.virt.hardware [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.084523] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c3fc2e-c5d8-4f03-972d-20bf3301100f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.092618] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5fda69-9bb5-42e1-b6ee-2c96665cb4e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.225440] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271659, 'name': Destroy_Task, 'duration_secs': 0.303595} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.225728] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Destroyed the VM [ 827.226185] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 827.227315] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-aec65ab4-3751-4043-a0d5-4db683bca255 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.234018] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 827.234018] env[62235]: value = "task-1271661" [ 827.234018] env[62235]: _type = "Task" [ 827.234018] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.242201] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271661, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.259824] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.260147] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.424095] env[62235]: DEBUG nova.compute.manager [req-bec1150b-f0b0-4a53-9bec-cfc1fe6b466b req-be4e6648-2039-476e-913c-6aaf7c11f726 service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Received event network-vif-deleted-6a7bb767-b47e-4a64-ae3c-f05483ec7cd9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.424223] env[62235]: INFO nova.compute.manager [req-bec1150b-f0b0-4a53-9bec-cfc1fe6b466b req-be4e6648-2039-476e-913c-6aaf7c11f726 service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Neutron deleted interface 6a7bb767-b47e-4a64-ae3c-f05483ec7cd9; detaching it from the instance and deleting it from the info cache [ 827.424400] env[62235]: DEBUG nova.network.neutron [req-bec1150b-f0b0-4a53-9bec-cfc1fe6b466b req-be4e6648-2039-476e-913c-6aaf7c11f726 service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.426569] env[62235]: DEBUG nova.scheduler.client.report [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.526861] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.284472} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.527128] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.527914] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b905b2-fed3-4924-8a82-9cfc5f4cdbb8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.551756] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 06ec6b9c-bef0-4f96-a185-4315961de7f7/06ec6b9c-bef0-4f96-a185-4315961de7f7.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.552116] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13dfd6ab-3ebc-47e6-b06e-4a7f5f87fd86 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.574509] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 827.574509] env[62235]: value = "task-1271662" [ 827.574509] env[62235]: _type = "Task" [ 827.574509] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.583170] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271662, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.629900] env[62235]: DEBUG nova.network.neutron [-] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.746205] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271661, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.762975] env[62235]: DEBUG nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.782135] env[62235]: DEBUG nova.compute.manager [req-7896c611-d0dc-433f-b967-d92e220d0b86 req-a6effdf2-4358-494c-9435-b7ca03a17b9c service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Received event network-vif-plugged-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.782395] env[62235]: DEBUG oslo_concurrency.lockutils [req-7896c611-d0dc-433f-b967-d92e220d0b86 req-a6effdf2-4358-494c-9435-b7ca03a17b9c service nova] Acquiring lock "80a803d1-c034-42b6-a15d-9fc3893953ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.782814] env[62235]: DEBUG oslo_concurrency.lockutils [req-7896c611-d0dc-433f-b967-d92e220d0b86 req-a6effdf2-4358-494c-9435-b7ca03a17b9c service nova] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.783659] env[62235]: DEBUG oslo_concurrency.lockutils [req-7896c611-d0dc-433f-b967-d92e220d0b86 req-a6effdf2-4358-494c-9435-b7ca03a17b9c service nova] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.783935] env[62235]: DEBUG nova.compute.manager [req-7896c611-d0dc-433f-b967-d92e220d0b86 req-a6effdf2-4358-494c-9435-b7ca03a17b9c service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] No waiting events found dispatching network-vif-plugged-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.784205] env[62235]: WARNING nova.compute.manager [req-7896c611-d0dc-433f-b967-d92e220d0b86 req-a6effdf2-4358-494c-9435-b7ca03a17b9c service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Received unexpected event network-vif-plugged-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 for instance with vm_state building and task_state spawning. [ 827.850901] env[62235]: INFO nova.compute.manager [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Rescuing [ 827.851126] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.851413] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquired lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.851659] env[62235]: DEBUG nova.network.neutron [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.929708] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "c571a362-9b55-42d3-a019-1695c1caf248" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.930099] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "c571a362-9b55-42d3-a019-1695c1caf248" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.931559] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8b62647-1b55-4148-b8a1-2922d0bb6a29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.935978] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.941047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.679s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.941047] env[62235]: INFO nova.compute.claims [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.951092] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928c379a-aefc-4e60-ae8a-f04f65d6e3d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.967226] env[62235]: INFO nova.scheduler.client.report [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleted allocations for instance b4a3777f-2ee5-464b-84eb-0722d59f2b44 [ 827.988579] env[62235]: DEBUG nova.compute.manager [req-bec1150b-f0b0-4a53-9bec-cfc1fe6b466b req-be4e6648-2039-476e-913c-6aaf7c11f726 service nova] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Detach interface failed, port_id=6a7bb767-b47e-4a64-ae3c-f05483ec7cd9, reason: Instance b5a1da09-2137-4434-a1a4-174b196b61b0 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 828.080243] env[62235]: DEBUG nova.network.neutron [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Successfully updated port: 5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.088644] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271662, 'name': ReconfigVM_Task, 'duration_secs': 0.325797} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.089282] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 06ec6b9c-bef0-4f96-a185-4315961de7f7/06ec6b9c-bef0-4f96-a185-4315961de7f7.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.090630] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3710d600-1e74-4cad-a844-25121f05d528 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.097698] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 828.097698] env[62235]: value = "task-1271663" [ 828.097698] env[62235]: _type = "Task" [ 828.097698] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.106048] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271663, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.133561] env[62235]: INFO nova.compute.manager [-] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Took 1.60 seconds to deallocate network for instance. [ 828.246050] env[62235]: DEBUG oslo_vmware.api [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271661, 'name': RemoveSnapshot_Task, 'duration_secs': 0.789184} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.246050] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 828.246277] env[62235]: INFO nova.compute.manager [None req-31b5c5cf-d18e-4282-9656-a928da049ac4 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Took 15.42 seconds to snapshot the instance on the hypervisor. [ 828.284250] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.440901] env[62235]: DEBUG nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.476976] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd9f6e37-2ae1-4ac3-9323-054d643ad34c tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "b4a3777f-2ee5-464b-84eb-0722d59f2b44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.604s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.560831] env[62235]: DEBUG nova.network.neutron [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [{"id": "600609cd-3b8f-4683-9136-308a989872ec", "address": "fa:16:3e:9f:b2:2e", "network": {"id": "e3892181-6713-45be-9572-f39a5a3955b2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-769806034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e340ac03d3734399a92abe4ed014251c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600609cd-3b", "ovs_interfaceid": "600609cd-3b8f-4683-9136-308a989872ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.589955] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.590133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquired lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.590286] env[62235]: DEBUG nova.network.neutron [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.607394] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271663, 'name': Rename_Task, 'duration_secs': 0.181274} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.607665] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.607906] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02a5aadf-78ff-43c2-ab39-2cd3d92b900b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.614227] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 828.614227] env[62235]: value = "task-1271664" [ 828.614227] env[62235]: _type = "Task" [ 828.614227] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.622172] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.641030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.963774] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.063379] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Releasing lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.131443] env[62235]: DEBUG oslo_vmware.api [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271664, 'name': PowerOnVM_Task, 'duration_secs': 0.462169} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.132614] env[62235]: DEBUG nova.network.neutron [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.135888] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.136224] env[62235]: INFO nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Took 8.33 seconds to spawn the instance on the hypervisor. [ 829.136521] env[62235]: DEBUG nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.140866] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8feeafc-71ec-4780-b8ef-f9b418556a6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.214472] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab63391-4313-4c7e-88b3-d7e0b74cc13d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.223137] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c09c363-e48c-476e-a691-6c9d4f74f056 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.258336] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d1c971-df25-4d76-9008-053de9ee4b8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.266914] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4415154c-11bf-497e-b472-7ff932dd8ccb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.281497] env[62235]: DEBUG nova.compute.provider_tree [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.326964] env[62235]: DEBUG nova.network.neutron [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Updating instance_info_cache with network_info: [{"id": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "address": "fa:16:3e:9a:2c:29", "network": {"id": "c12598e4-11de-43b5-a569-b030b45411a7", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1353190335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940941e982494a4caeebda900b3e0b8c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f04dc27-a2", "ovs_interfaceid": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.597995] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 829.598292] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a2545a4-ffc2-4d8b-a336-928f8e7ed944 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.606555] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 829.606555] env[62235]: value = "task-1271665" [ 829.606555] env[62235]: _type = "Task" [ 829.606555] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.614620] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.631429] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "ec1830c7-7095-45d1-a188-c2993c60035e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.631594] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "ec1830c7-7095-45d1-a188-c2993c60035e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.631875] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "ec1830c7-7095-45d1-a188-c2993c60035e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.632136] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "ec1830c7-7095-45d1-a188-c2993c60035e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.632344] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "ec1830c7-7095-45d1-a188-c2993c60035e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.634751] env[62235]: INFO nova.compute.manager [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Terminating instance [ 829.641021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "refresh_cache-ec1830c7-7095-45d1-a188-c2993c60035e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.641021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquired lock "refresh_cache-ec1830c7-7095-45d1-a188-c2993c60035e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.641021] env[62235]: DEBUG nova.network.neutron [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.662897] env[62235]: INFO nova.compute.manager [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Took 42.35 seconds to build instance. [ 829.784262] env[62235]: DEBUG nova.scheduler.client.report [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.829491] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Releasing lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.829793] env[62235]: DEBUG nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Instance network_info: |[{"id": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "address": "fa:16:3e:9a:2c:29", "network": {"id": "c12598e4-11de-43b5-a569-b030b45411a7", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1353190335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940941e982494a4caeebda900b3e0b8c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f04dc27-a2", "ovs_interfaceid": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.830250] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:2c:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94926d5b-bfab-4c04-85b5-0fe89934c8ff', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f04dc27-a2d4-46cc-bf3c-947c4c3dab59', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.837620] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Creating folder: Project (940941e982494a4caeebda900b3e0b8c). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.838213] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96148a2b-86f3-427b-9045-1c4aea122531 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.849185] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Created folder: Project (940941e982494a4caeebda900b3e0b8c) in parent group-v273362. [ 829.849380] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Creating folder: Instances. Parent ref: group-v273433. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.849616] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0929a6ff-e0c9-43e2-bc2c-4f29953600eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.858715] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Created folder: Instances in parent group-v273433. [ 829.858938] env[62235]: DEBUG oslo.service.loopingcall [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.859145] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.859349] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70bcb710-15b9-4d0e-b65d-d4026258cbc0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.878087] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.878087] env[62235]: value = "task-1271668" [ 829.878087] env[62235]: _type = "Task" [ 829.878087] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.886763] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271668, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.952509] env[62235]: DEBUG nova.compute.manager [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Received event network-changed-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.952509] env[62235]: DEBUG nova.compute.manager [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Refreshing instance network info cache due to event network-changed-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.952509] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] Acquiring lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.952509] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] Acquired lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.952509] env[62235]: DEBUG nova.network.neutron [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Refreshing network info cache for port 5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.117081] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271665, 'name': PowerOffVM_Task, 'duration_secs': 0.340173} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.117382] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.118189] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca69611-1729-4b6c-adab-4c1b3155feb1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.136607] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e597e4-76cc-46ad-8418-f91d3f5f1efb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.158216] env[62235]: DEBUG nova.network.neutron [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.166076] env[62235]: DEBUG oslo_concurrency.lockutils [None req-457b7f1a-33a8-441f-a11d-b581b12e41d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.892s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.175558] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.175888] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb91be39-8f99-4848-8145-becad3e7d0b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.184468] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 830.184468] env[62235]: value = "task-1271669" [ 830.184468] env[62235]: _type = "Task" [ 830.184468] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.192694] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.230625] env[62235]: DEBUG nova.network.neutron [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.289643] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.290193] env[62235]: DEBUG nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.292960] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.549s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.294963] env[62235]: INFO nova.compute.claims [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.388642] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271668, 'name': CreateVM_Task, 'duration_secs': 0.344468} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.389757] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.389757] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.389891] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.390748] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.390748] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bcbd98f-3304-4f2e-846e-91e4a452708a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.395178] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 830.395178] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5260316b-f502-e6cb-a34a-2c0c3f0af0ad" [ 830.395178] env[62235]: _type = "Task" [ 830.395178] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.402960] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5260316b-f502-e6cb-a34a-2c0c3f0af0ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.682914] env[62235]: DEBUG nova.network.neutron [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Updated VIF entry in instance network info cache for port 5f04dc27-a2d4-46cc-bf3c-947c4c3dab59. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 830.683290] env[62235]: DEBUG nova.network.neutron [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Updating instance_info_cache with network_info: [{"id": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "address": "fa:16:3e:9a:2c:29", "network": {"id": "c12598e4-11de-43b5-a569-b030b45411a7", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1353190335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940941e982494a4caeebda900b3e0b8c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f04dc27-a2", "ovs_interfaceid": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.698175] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 830.698404] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.698808] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.698996] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.699203] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.699703] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84a1446a-cb79-437c-8abb-008e5c065295 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.710110] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.710997] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.710997] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4717b0a3-131c-4d65-90f4-17c020eda987 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.716474] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 830.716474] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528d9e97-4595-95cd-e64f-69bbdd7741c8" [ 830.716474] env[62235]: _type = "Task" [ 830.716474] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.724304] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528d9e97-4595-95cd-e64f-69bbdd7741c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.732898] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Releasing lock "refresh_cache-ec1830c7-7095-45d1-a188-c2993c60035e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.733314] env[62235]: DEBUG nova.compute.manager [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 830.733512] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 830.734627] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8e423e-da17-47c1-b2dd-ba2ae9590257 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.741390] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.741614] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9fc89d2-a7e0-41ee-98e6-d477ed82f1b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.747811] env[62235]: DEBUG oslo_vmware.api [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 830.747811] env[62235]: value = "task-1271670" [ 830.747811] env[62235]: _type = "Task" [ 830.747811] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.755071] env[62235]: DEBUG oslo_vmware.api [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.799392] env[62235]: DEBUG nova.compute.utils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.802839] env[62235]: DEBUG nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.803030] env[62235]: DEBUG nova.network.neutron [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.861274] env[62235]: DEBUG nova.policy [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b64f6d6696c042079cced4c83d161f6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4469550d32b4482a9d5e9244cbd681e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.905306] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5260316b-f502-e6cb-a34a-2c0c3f0af0ad, 'name': SearchDatastore_Task, 'duration_secs': 0.014185} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.905629] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.905875] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.906107] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.149237] env[62235]: DEBUG nova.network.neutron [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Successfully created port: bfd2c6bf-a775-49d3-8e77-b851443ec66b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.185965] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b2d4cb2-44ca-4bc7-b63e-3e2ddbdde4e6 req-ae2586f2-2141-4945-9360-5f8082a6fa45 service nova] Releasing lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.227220] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528d9e97-4595-95cd-e64f-69bbdd7741c8, 'name': SearchDatastore_Task, 'duration_secs': 0.009617} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.228046] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6327035-1f99-423e-9547-26c2c75a60d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.234666] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 831.234666] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f55e73-16f8-c8e0-8888-8a4afe3d0b1d" [ 831.234666] env[62235]: _type = "Task" [ 831.234666] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.242419] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f55e73-16f8-c8e0-8888-8a4afe3d0b1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.256978] env[62235]: DEBUG oslo_vmware.api [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271670, 'name': PowerOffVM_Task, 'duration_secs': 0.170605} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.257259] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 831.257471] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 831.257725] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4b15b4e-70ec-41c0-9743-a487b24de2db {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.287371] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 831.287616] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 831.287807] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleting the datastore file [datastore1] ec1830c7-7095-45d1-a188-c2993c60035e {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.288105] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de91ce27-9ef5-43a7-9790-2c6bbb3c2e88 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.295865] env[62235]: DEBUG oslo_vmware.api [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for the task: (returnval){ [ 831.295865] env[62235]: value = "task-1271672" [ 831.295865] env[62235]: _type = "Task" [ 831.295865] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.304749] env[62235]: DEBUG nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.311017] env[62235]: DEBUG oslo_vmware.api [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271672, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.533965] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142dcfa4-08fb-47c1-9c19-9d36f72873cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.542215] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a763f07-edec-4994-a2f7-d7aba8d2883c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.575720] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e048257-7940-4039-b6f2-302e9985161d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.584930] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd75a47d-7957-47dc-8db8-b5dbdfd7bfb5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.599864] env[62235]: DEBUG nova.compute.provider_tree [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.745929] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f55e73-16f8-c8e0-8888-8a4afe3d0b1d, 'name': SearchDatastore_Task, 'duration_secs': 0.010415} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.746288] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.746577] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk. {{(pid=62235) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 831.746904] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.747128] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.747375] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54528e3c-a2cc-4c6c-b039-7fa19719625c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.749578] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9dedc103-c34a-46c5-83e2-9ee9b84450aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.758884] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 831.758884] env[62235]: value = "task-1271673" [ 831.758884] env[62235]: _type = "Task" [ 831.758884] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.760286] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.760286] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.764025] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73b0c261-a052-465a-9b85-a18e444b8c8b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.770788] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 831.770788] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52288847-d73b-becb-32c7-21035308580d" [ 831.770788] env[62235]: _type = "Task" [ 831.770788] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.774833] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.784247] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52288847-d73b-becb-32c7-21035308580d, 'name': SearchDatastore_Task, 'duration_secs': 0.010212} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.785092] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9fafc41-d5c5-4f9a-892a-2072ea0a1bcc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.795027] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 831.795027] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e453f9-2a83-11f9-3c5d-5e7330de77b6" [ 831.795027] env[62235]: _type = "Task" [ 831.795027] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.806720] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e453f9-2a83-11f9-3c5d-5e7330de77b6, 'name': SearchDatastore_Task, 'duration_secs': 0.010855} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.809861] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.810153] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 80a803d1-c034-42b6-a15d-9fc3893953ad/80a803d1-c034-42b6-a15d-9fc3893953ad.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.813599] env[62235]: DEBUG oslo_vmware.api [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Task: {'id': task-1271672, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104574} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.814029] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a8d46d3-33b8-46fb-8959-d22feac349b4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.816068] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.816295] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.816488] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.816670] env[62235]: INFO nova.compute.manager [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Took 1.08 seconds to destroy the instance on the hypervisor. [ 831.816910] env[62235]: DEBUG oslo.service.loopingcall [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.817483] env[62235]: DEBUG nova.compute.manager [-] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 831.817483] env[62235]: DEBUG nova.network.neutron [-] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 831.825275] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 831.825275] env[62235]: value = "task-1271674" [ 831.825275] env[62235]: _type = "Task" [ 831.825275] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.835278] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271674, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.836562] env[62235]: DEBUG nova.network.neutron [-] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.983262] env[62235]: DEBUG nova.compute.manager [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Received event network-changed-0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.983484] env[62235]: DEBUG nova.compute.manager [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Refreshing instance network info cache due to event network-changed-0d781479-5511-42b0-9aa5-4efbe57d8d58. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.983717] env[62235]: DEBUG oslo_concurrency.lockutils [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] Acquiring lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.983868] env[62235]: DEBUG oslo_concurrency.lockutils [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] Acquired lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.984044] env[62235]: DEBUG nova.network.neutron [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Refreshing network info cache for port 0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.102862] env[62235]: DEBUG nova.scheduler.client.report [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.269915] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502156} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.270260] env[62235]: INFO nova.virt.vmwareapi.ds_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk. [ 832.271103] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb68588-f4a5-4a66-b1f3-dd97d831d914 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.298948] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.299310] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46d05bf5-9f12-4de4-bd07-36644a913d45 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.315372] env[62235]: DEBUG nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.321102] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 832.321102] env[62235]: value = "task-1271675" [ 832.321102] env[62235]: _type = "Task" [ 832.321102] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.334589] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271674, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.338398] env[62235]: DEBUG nova.network.neutron [-] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.339736] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271675, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.348871] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.349168] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.349339] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.349537] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.349692] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.349848] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.350072] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.350246] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.350429] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.350655] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.350792] env[62235]: DEBUG nova.virt.hardware [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.351933] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed3bfd5-8d0e-4160-89bc-02d2ff22aca8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.361913] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a15feb7-9792-41b3-9fe8-0a85f3b94976 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.608949] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.609567] env[62235]: DEBUG nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.612117] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.808s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.613903] env[62235]: INFO nova.compute.claims [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.711607] env[62235]: DEBUG nova.network.neutron [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Successfully updated port: bfd2c6bf-a775-49d3-8e77-b851443ec66b {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.745893] env[62235]: DEBUG nova.network.neutron [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updated VIF entry in instance network info cache for port 0d781479-5511-42b0-9aa5-4efbe57d8d58. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.746576] env[62235]: DEBUG nova.network.neutron [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updating instance_info_cache with network_info: [{"id": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "address": "fa:16:3e:03:4a:67", "network": {"id": "df25d9d9-5721-4c36-b75a-2cf35c302b03", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-520889908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f2551ad3cfb4a0d8112f85d4e95fe64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d781479-55", "ovs_interfaceid": "0d781479-5511-42b0-9aa5-4efbe57d8d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.832686] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271675, 'name': ReconfigVM_Task, 'duration_secs': 0.474237} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.833498] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.834435] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472b32f1-10e6-468f-97bc-512bdaf5b781 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.839775] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271674, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.747409} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.840399] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 80a803d1-c034-42b6-a15d-9fc3893953ad/80a803d1-c034-42b6-a15d-9fc3893953ad.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.840575] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.841032] env[62235]: INFO nova.compute.manager [-] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Took 1.02 seconds to deallocate network for instance. [ 832.841295] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-140c2d18-40b8-4075-b9b4-d8b9b41780e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.869664] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86a642d3-bd62-41dc-a66d-c192dd7959b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.882468] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 832.882468] env[62235]: value = "task-1271676" [ 832.882468] env[62235]: _type = "Task" [ 832.882468] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.889256] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 832.889256] env[62235]: value = "task-1271677" [ 832.889256] env[62235]: _type = "Task" [ 832.889256] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.892750] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271676, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.901396] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271677, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.122845] env[62235]: DEBUG nova.compute.utils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.124768] env[62235]: DEBUG nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.124768] env[62235]: DEBUG nova.network.neutron [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.174232] env[62235]: DEBUG nova.policy [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f6f9cd7af6784e468ff9e3230fdb433f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d1081d5b610424d833a5f7611c1e796', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.213954] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-00825c9e-b398-4edc-a5ed-e4b5b476af1c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.214117] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-00825c9e-b398-4edc-a5ed-e4b5b476af1c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.214282] env[62235]: DEBUG nova.network.neutron [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.249255] env[62235]: DEBUG oslo_concurrency.lockutils [req-e63b3ebf-cadb-42d6-93e2-615b1be6db5a req-6ccec99f-0fe8-4c58-93fd-303a106a5d0a service nova] Releasing lock "refresh_cache-06ec6b9c-bef0-4f96-a185-4315961de7f7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.382454] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.394355] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271676, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.256279} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.397702] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.398538] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3cf654b-a2e9-446a-b8de-0f905f8323fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.426869] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 80a803d1-c034-42b6-a15d-9fc3893953ad/80a803d1-c034-42b6-a15d-9fc3893953ad.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.427284] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271677, 'name': ReconfigVM_Task, 'duration_secs': 0.166291} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.427506] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05a8ffc9-7778-488f-bba3-2b50c997ae31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.441892] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.442806] env[62235]: DEBUG nova.network.neutron [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Successfully created port: 5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.444609] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64180c16-e1a8-4c6a-92c3-9b0ec14c375d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.453023] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 833.453023] env[62235]: value = "task-1271678" [ 833.453023] env[62235]: _type = "Task" [ 833.453023] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.454120] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 833.454120] env[62235]: value = "task-1271679" [ 833.454120] env[62235]: _type = "Task" [ 833.454120] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.467132] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271679, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.470130] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271678, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.631423] env[62235]: DEBUG nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.762221] env[62235]: DEBUG nova.network.neutron [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.881605] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ecbe2e-5385-4750-bae2-8c8d6c8ae12a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.894196] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0eee3d3-001b-4ede-badc-a35bb109d0a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.926604] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce852806-e574-47d6-9ebe-035a6b42681a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.934866] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b769d016-ed09-451f-b73a-e2dcd81f101e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.940259] env[62235]: DEBUG nova.network.neutron [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Updating instance_info_cache with network_info: [{"id": "bfd2c6bf-a775-49d3-8e77-b851443ec66b", "address": "fa:16:3e:2c:c1:e8", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd2c6bf-a7", "ovs_interfaceid": "bfd2c6bf-a775-49d3-8e77-b851443ec66b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.949541] env[62235]: DEBUG nova.compute.provider_tree [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 833.969630] env[62235]: DEBUG oslo_vmware.api [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271678, 'name': PowerOnVM_Task, 'duration_secs': 0.409097} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.973474] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.976068] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271679, 'name': ReconfigVM_Task, 'duration_secs': 0.342314} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.978232] env[62235]: DEBUG nova.compute.manager [None req-fe9eb5e9-b216-4252-9cbf-3106a3d89696 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.978232] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 80a803d1-c034-42b6-a15d-9fc3893953ad/80a803d1-c034-42b6-a15d-9fc3893953ad.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.978730] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f679fa55-744d-4e46-a9b5-fc01d3d5574e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.982679] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a2660f4-490b-45fd-bc98-84a2b0918ae7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.994627] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 833.994627] env[62235]: value = "task-1271680" [ 833.994627] env[62235]: _type = "Task" [ 833.994627] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.003780] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271680, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.045748] env[62235]: DEBUG nova.compute.manager [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Received event network-vif-plugged-bfd2c6bf-a775-49d3-8e77-b851443ec66b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.045985] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] Acquiring lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.046214] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.046422] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.046943] env[62235]: DEBUG nova.compute.manager [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] No waiting events found dispatching network-vif-plugged-bfd2c6bf-a775-49d3-8e77-b851443ec66b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 834.046943] env[62235]: WARNING nova.compute.manager [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Received unexpected event network-vif-plugged-bfd2c6bf-a775-49d3-8e77-b851443ec66b for instance with vm_state building and task_state spawning. [ 834.046943] env[62235]: DEBUG nova.compute.manager [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Received event network-changed-bfd2c6bf-a775-49d3-8e77-b851443ec66b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.047109] env[62235]: DEBUG nova.compute.manager [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Refreshing instance network info cache due to event network-changed-bfd2c6bf-a775-49d3-8e77-b851443ec66b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 834.047260] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] Acquiring lock "refresh_cache-00825c9e-b398-4edc-a5ed-e4b5b476af1c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.451824] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-00825c9e-b398-4edc-a5ed-e4b5b476af1c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.452188] env[62235]: DEBUG nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Instance network_info: |[{"id": "bfd2c6bf-a775-49d3-8e77-b851443ec66b", "address": "fa:16:3e:2c:c1:e8", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd2c6bf-a7", "ovs_interfaceid": "bfd2c6bf-a775-49d3-8e77-b851443ec66b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.455884] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] Acquired lock "refresh_cache-00825c9e-b398-4edc-a5ed-e4b5b476af1c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.456121] env[62235]: DEBUG nova.network.neutron [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Refreshing network info cache for port bfd2c6bf-a775-49d3-8e77-b851443ec66b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.457498] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:c1:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfd2c6bf-a775-49d3-8e77-b851443ec66b', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.464920] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Creating folder: Project (4469550d32b4482a9d5e9244cbd681e1). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.465905] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4209950c-28b4-43ae-8ff9-74e9e55e23d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.470394] env[62235]: ERROR nova.scheduler.client.report [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [req-7acd7f9e-9945-42ec-b1fa-a3a77cb2c85f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7acd7f9e-9945-42ec-b1fa-a3a77cb2c85f"}]} [ 834.481700] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Created folder: Project (4469550d32b4482a9d5e9244cbd681e1) in parent group-v273362. [ 834.481899] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Creating folder: Instances. Parent ref: group-v273436. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 834.482152] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf7306a0-9b6c-4c81-84f0-211f18d8cca4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.485971] env[62235]: DEBUG nova.scheduler.client.report [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 834.491881] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Created folder: Instances in parent group-v273436. [ 834.492134] env[62235]: DEBUG oslo.service.loopingcall [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.492596] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 834.492596] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39d6436d-a275-4510-88f4-bd3b637278ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.507268] env[62235]: DEBUG nova.scheduler.client.report [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 834.507572] env[62235]: DEBUG nova.compute.provider_tree [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.521270] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271680, 'name': Rename_Task, 'duration_secs': 0.145501} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.523106] env[62235]: DEBUG nova.scheduler.client.report [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 834.525044] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 834.525281] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.525281] env[62235]: value = "task-1271683" [ 834.525281] env[62235]: _type = "Task" [ 834.525281] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.526195] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-398d4aa3-b4a2-4d4a-a72d-6862d44eb0dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.537805] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271683, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.539236] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 834.539236] env[62235]: value = "task-1271684" [ 834.539236] env[62235]: _type = "Task" [ 834.539236] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.543252] env[62235]: DEBUG nova.scheduler.client.report [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 834.551627] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271684, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.638420] env[62235]: DEBUG nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.666187] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.666464] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.666690] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.666819] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.666970] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.668814] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.669088] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.669271] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.669450] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.669624] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.669802] env[62235]: DEBUG nova.virt.hardware [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.670714] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646aeb17-fc5f-4d41-96c6-19e38ce57e6a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.681392] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59828b4c-9a14-4f16-a958-96aa95529572 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.805357] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42536ce-0f46-42c6-b178-bf21e2f3faa0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.813419] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb00050-124a-4f6c-8a2e-db21af84d17b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.854287] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb45291-8e87-4777-82b5-1a680ae94d16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.862845] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60baeeb-5faf-4daa-9641-4f536fc67148 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.872364] env[62235]: DEBUG nova.compute.manager [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received event network-changed-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.872573] env[62235]: DEBUG nova.compute.manager [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing instance network info cache due to event network-changed-600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 834.872838] env[62235]: DEBUG oslo_concurrency.lockutils [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] Acquiring lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.873101] env[62235]: DEBUG oslo_concurrency.lockutils [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] Acquired lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.873358] env[62235]: DEBUG nova.network.neutron [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing network info cache for port 600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.884303] env[62235]: DEBUG nova.compute.provider_tree [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.038767] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271683, 'name': CreateVM_Task, 'duration_secs': 0.365743} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.039028] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 835.039788] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.039966] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.040353] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.044019] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ff6a419-2092-4d9c-b6d0-00af44e20a1b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.049502] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 835.049502] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e35d86-14e2-366d-9ebf-0f46b515d078" [ 835.049502] env[62235]: _type = "Task" [ 835.049502] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.053219] env[62235]: DEBUG oslo_vmware.api [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271684, 'name': PowerOnVM_Task, 'duration_secs': 0.470484} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.056222] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 835.056501] env[62235]: INFO nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Took 8.00 seconds to spawn the instance on the hypervisor. [ 835.056681] env[62235]: DEBUG nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.057548] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a877881-96f1-46a3-b4cb-a6c7175b1143 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.070910] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e35d86-14e2-366d-9ebf-0f46b515d078, 'name': SearchDatastore_Task, 'duration_secs': 0.011706} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.071404] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.071785] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.072059] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.072214] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.072397] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.074590] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99b31dae-dc81-4db5-ab7e-313ae1950a12 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.083080] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.083299] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.084126] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1aadf972-d8cc-4dd5-9144-70c5e3631821 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.090495] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 835.090495] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c29db5-0aa1-aa42-865b-5c1ac3c0e7f2" [ 835.090495] env[62235]: _type = "Task" [ 835.090495] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.103329] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c29db5-0aa1-aa42-865b-5c1ac3c0e7f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.157088] env[62235]: DEBUG nova.network.neutron [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Successfully updated port: 5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.331495] env[62235]: DEBUG nova.network.neutron [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Updated VIF entry in instance network info cache for port bfd2c6bf-a775-49d3-8e77-b851443ec66b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.332048] env[62235]: DEBUG nova.network.neutron [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Updating instance_info_cache with network_info: [{"id": "bfd2c6bf-a775-49d3-8e77-b851443ec66b", "address": "fa:16:3e:2c:c1:e8", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd2c6bf-a7", "ovs_interfaceid": "bfd2c6bf-a775-49d3-8e77-b851443ec66b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.430815] env[62235]: DEBUG nova.scheduler.client.report [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Updated inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 835.430815] env[62235]: DEBUG nova.compute.provider_tree [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 81 to 82 during operation: update_inventory {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 835.430815] env[62235]: DEBUG nova.compute.provider_tree [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.584447] env[62235]: INFO nova.compute.manager [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Took 40.63 seconds to build instance. [ 835.602917] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c29db5-0aa1-aa42-865b-5c1ac3c0e7f2, 'name': SearchDatastore_Task, 'duration_secs': 0.009151} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.603716] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a298bc82-5730-4b4f-9f65-835bb8c9471b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.610035] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 835.610035] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a329f5-dc93-d0f1-4234-7cf6bf94f59a" [ 835.610035] env[62235]: _type = "Task" [ 835.610035] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.618941] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a329f5-dc93-d0f1-4234-7cf6bf94f59a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.641926] env[62235]: DEBUG nova.network.neutron [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updated VIF entry in instance network info cache for port 600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.642114] env[62235]: DEBUG nova.network.neutron [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [{"id": "600609cd-3b8f-4683-9136-308a989872ec", "address": "fa:16:3e:9f:b2:2e", "network": {"id": "e3892181-6713-45be-9572-f39a5a3955b2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-769806034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e340ac03d3734399a92abe4ed014251c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600609cd-3b", "ovs_interfaceid": "600609cd-3b8f-4683-9136-308a989872ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.658818] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.658958] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquired lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.659106] env[62235]: DEBUG nova.network.neutron [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.835174] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c40d327-8094-47bc-88cd-d6cb8e411aae req-a0b5e9d9-c46d-4d03-b79e-01d3d716709b service nova] Releasing lock "refresh_cache-00825c9e-b398-4edc-a5ed-e4b5b476af1c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.935695] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.323s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.936228] env[62235]: DEBUG nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.938926] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.940s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.939164] env[62235]: DEBUG nova.objects.instance [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lazy-loading 'resources' on Instance uuid 590440de-4b3b-4019-b0c3-58b5ad329c88 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 836.077429] env[62235]: DEBUG nova.compute.manager [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received event network-changed-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.077593] env[62235]: DEBUG nova.compute.manager [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing instance network info cache due to event network-changed-600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.077900] env[62235]: DEBUG oslo_concurrency.lockutils [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] Acquiring lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.084174] env[62235]: DEBUG oslo_concurrency.lockutils [None req-88066207-9d7a-493b-835c-c1d57392c071 tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.817s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.121817] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a329f5-dc93-d0f1-4234-7cf6bf94f59a, 'name': SearchDatastore_Task, 'duration_secs': 0.009645} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.122112] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.122384] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.122757] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97869ef3-ce56-4f07-b320-4bde6ec53c0d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.131846] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 836.131846] env[62235]: value = "task-1271685" [ 836.131846] env[62235]: _type = "Task" [ 836.131846] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.140723] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.144819] env[62235]: DEBUG oslo_concurrency.lockutils [req-7bbf09bc-f9af-4693-8fc1-54c803890d19 req-14dbeeee-0831-49e2-8722-10b649d37e3f service nova] Releasing lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.145165] env[62235]: DEBUG oslo_concurrency.lockutils [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] Acquired lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.145368] env[62235]: DEBUG nova.network.neutron [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing network info cache for port 600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.205890] env[62235]: DEBUG nova.network.neutron [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.386222] env[62235]: DEBUG nova.network.neutron [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updating instance_info_cache with network_info: [{"id": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "address": "fa:16:3e:3c:21:f8", "network": {"id": "bcc80e56-2152-4dd0-829f-14cc48fd0e88", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1093381308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d1081d5b610424d833a5f7611c1e796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba39fe2-6d", "ovs_interfaceid": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.446055] env[62235]: DEBUG nova.compute.utils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.448785] env[62235]: DEBUG nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.448987] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 836.572186] env[62235]: DEBUG nova.policy [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5ec091a9504485bba6048fce1a1d090', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91d149e97a7b4bdfa43ed32cb59d0fdf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.643181] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481755} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.646166] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.646696] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.647257] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49b5b4e8-f200-49d6-82d7-a4cd08e112e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.659177] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 836.659177] env[62235]: value = "task-1271686" [ 836.659177] env[62235]: _type = "Task" [ 836.659177] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.670085] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271686, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.751018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed093295-fc88-4b07-8c21-795fc157320f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.758512] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc5929f-8747-45cb-a986-272965421901 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.792764] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142b0159-46e2-48c8-a427-fce7ff0093a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.802729] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6652180-277d-4fa2-a7e6-b18e7b10fab7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.823018] env[62235]: DEBUG nova.compute.provider_tree [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.891469] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Releasing lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.891940] env[62235]: DEBUG nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Instance network_info: |[{"id": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "address": "fa:16:3e:3c:21:f8", "network": {"id": "bcc80e56-2152-4dd0-829f-14cc48fd0e88", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1093381308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d1081d5b610424d833a5f7611c1e796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba39fe2-6d", "ovs_interfaceid": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.892303] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:21:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e028024-a9c1-4cae-8849-ea770a7ae0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ba39fe2-6d0a-4a24-8b3f-024002027e22', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.902518] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Creating folder: Project (6d1081d5b610424d833a5f7611c1e796). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.906911] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-31cb563f-b92f-4aab-92ad-0cc48d6a3dce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.925368] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Created folder: Project (6d1081d5b610424d833a5f7611c1e796) in parent group-v273362. [ 836.925969] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Creating folder: Instances. Parent ref: group-v273439. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.926361] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f820bc79-ccdc-4bd6-84dc-90a573f87947 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.942496] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Created folder: Instances in parent group-v273439. [ 836.943012] env[62235]: DEBUG oslo.service.loopingcall [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.943566] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.943566] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcfca46d-9b2f-4ae5-ba27-0f904aa3fdf9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.963466] env[62235]: DEBUG nova.network.neutron [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updated VIF entry in instance network info cache for port 600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.964108] env[62235]: DEBUG nova.network.neutron [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [{"id": "600609cd-3b8f-4683-9136-308a989872ec", "address": "fa:16:3e:9f:b2:2e", "network": {"id": "e3892181-6713-45be-9572-f39a5a3955b2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-769806034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e340ac03d3734399a92abe4ed014251c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600609cd-3b", "ovs_interfaceid": "600609cd-3b8f-4683-9136-308a989872ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.965456] env[62235]: DEBUG nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.976509] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.976509] env[62235]: value = "task-1271689" [ 836.976509] env[62235]: _type = "Task" [ 836.976509] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.986987] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271689, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.175197] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271686, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062664} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.175560] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.176301] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f917c56-994d-4ed6-b96e-aa16daacf54a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.200811] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.201127] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d810334d-9a7e-4b6e-a9cf-ae21ac8c1c02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.221590] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 837.221590] env[62235]: value = "task-1271690" [ 837.221590] env[62235]: _type = "Task" [ 837.221590] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.230061] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271690, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.327187] env[62235]: DEBUG nova.scheduler.client.report [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.333799] env[62235]: DEBUG nova.compute.manager [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Received event network-vif-plugged-5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.334070] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Acquiring lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.334252] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.334489] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.334654] env[62235]: DEBUG nova.compute.manager [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] No waiting events found dispatching network-vif-plugged-5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 837.334831] env[62235]: WARNING nova.compute.manager [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Received unexpected event network-vif-plugged-5ba39fe2-6d0a-4a24-8b3f-024002027e22 for instance with vm_state building and task_state spawning. [ 837.335061] env[62235]: DEBUG nova.compute.manager [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Received event network-changed-5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.335173] env[62235]: DEBUG nova.compute.manager [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Refreshing instance network info cache due to event network-changed-5ba39fe2-6d0a-4a24-8b3f-024002027e22. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.335364] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Acquiring lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.335539] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Acquired lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.335673] env[62235]: DEBUG nova.network.neutron [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Refreshing network info cache for port 5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.448132] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Successfully created port: eede8be1-1cbc-4791-802e-52193488eb27 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.469456] env[62235]: DEBUG oslo_concurrency.lockutils [req-1f7673af-c027-44f1-938a-ae7c8c408742 req-965982a9-aa05-4544-be56-e9037c1fb133 service nova] Releasing lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.487122] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271689, 'name': CreateVM_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.735298] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271690, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.832953] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.835331] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.330s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.839788] env[62235]: INFO nova.compute.claims [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.865449] env[62235]: INFO nova.scheduler.client.report [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Deleted allocations for instance 590440de-4b3b-4019-b0c3-58b5ad329c88 [ 837.882604] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Successfully created port: 90607cc9-dfc7-4f25-b1dd-ad81a02691f9 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.979859] env[62235]: DEBUG nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.992181] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271689, 'name': CreateVM_Task, 'duration_secs': 0.607248} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.992360] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.993100] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.993310] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.993660] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.993958] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7587afbf-2591-4e95-b85d-ccee95059105 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.998952] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 837.998952] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b7fcb1-3ae1-7f70-29a2-dfed5047925c" [ 837.998952] env[62235]: _type = "Task" [ 837.998952] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.006924] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b7fcb1-3ae1-7f70-29a2-dfed5047925c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.018173] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.018428] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.018591] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.018783] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.018935] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.019129] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.019327] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.019495] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.019668] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.019852] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.020047] env[62235]: DEBUG nova.virt.hardware [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.021226] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd95a17-04c8-4fc1-9584-78b5e2d3d458 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.029421] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bd06dc-5a88-4bb8-b1bd-fbb453ea1904 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.236405] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271690, 'name': ReconfigVM_Task, 'duration_secs': 0.564159} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.236916] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.237737] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f0065d4-0c05-4f14-83be-f7344bb0cc50 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.245187] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 838.245187] env[62235]: value = "task-1271691" [ 838.245187] env[62235]: _type = "Task" [ 838.245187] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.257655] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271691, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.380223] env[62235]: DEBUG oslo_concurrency.lockutils [None req-008ae879-6289-4129-ab9f-4dbf292f8253 tempest-ServerAddressesNegativeTestJSON-476815878 tempest-ServerAddressesNegativeTestJSON-476815878-project-member] Lock "590440de-4b3b-4019-b0c3-58b5ad329c88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.194s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.392208] env[62235]: DEBUG nova.compute.manager [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Received event network-changed-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.395087] env[62235]: DEBUG nova.compute.manager [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Refreshing instance network info cache due to event network-changed-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.395087] env[62235]: DEBUG oslo_concurrency.lockutils [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] Acquiring lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.395087] env[62235]: DEBUG oslo_concurrency.lockutils [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] Acquired lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.395087] env[62235]: DEBUG nova.network.neutron [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Refreshing network info cache for port 5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.419983] env[62235]: DEBUG nova.network.neutron [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updated VIF entry in instance network info cache for port 5ba39fe2-6d0a-4a24-8b3f-024002027e22. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.420260] env[62235]: DEBUG nova.network.neutron [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updating instance_info_cache with network_info: [{"id": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "address": "fa:16:3e:3c:21:f8", "network": {"id": "bcc80e56-2152-4dd0-829f-14cc48fd0e88", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1093381308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d1081d5b610424d833a5f7611c1e796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba39fe2-6d", "ovs_interfaceid": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.512527] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b7fcb1-3ae1-7f70-29a2-dfed5047925c, 'name': SearchDatastore_Task, 'duration_secs': 0.029745} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.513075] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.513476] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.513735] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.514068] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.514295] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.514623] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a2f712d-2d50-4065-9dbc-347e590c6605 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.523123] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.523330] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.526914] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fc2c481-880b-4d8b-b25e-49ca9f14451b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.531343] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 838.531343] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52711899-e715-55b6-b818-5ea223c471ed" [ 838.531343] env[62235]: _type = "Task" [ 838.531343] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.541105] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52711899-e715-55b6-b818-5ea223c471ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.756280] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271691, 'name': Rename_Task, 'duration_secs': 0.333281} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.756595] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.756850] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d90a89c3-9613-4b32-9202-1e516654b4e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.763199] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 838.763199] env[62235]: value = "task-1271692" [ 838.763199] env[62235]: _type = "Task" [ 838.763199] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.770840] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.924752] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Releasing lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.925045] env[62235]: DEBUG nova.compute.manager [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received event network-changed-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.925176] env[62235]: DEBUG nova.compute.manager [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing instance network info cache due to event network-changed-600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.925426] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Acquiring lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.925770] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Acquired lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.925770] env[62235]: DEBUG nova.network.neutron [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing network info cache for port 600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 839.047931] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52711899-e715-55b6-b818-5ea223c471ed, 'name': SearchDatastore_Task, 'duration_secs': 0.018077} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.049071] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be7cd943-1072-43aa-88f6-75cf58299ca2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.053676] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 839.053676] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a2438-0270-1834-fa33-b945a315d7e3" [ 839.053676] env[62235]: _type = "Task" [ 839.053676] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.067032] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a2438-0270-1834-fa33-b945a315d7e3, 'name': SearchDatastore_Task, 'duration_secs': 0.009581} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.067032] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.067184] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9/9f5fcb83-bb6f-4815-b842-a37b2dbf26e9.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.067393] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53fa0fdc-8eb4-48ad-a444-660323227f54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.075762] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 839.075762] env[62235]: value = "task-1271693" [ 839.075762] env[62235]: _type = "Task" [ 839.075762] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.090495] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271693, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.112601] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8bd8c4-cdee-47e9-bed2-368e176ab820 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.119889] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24550b2-febc-49b0-b460-1b8c7b56682e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.153427] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b95db0c-6986-4ee8-be4c-1dd14ad79eaa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.160920] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3131848d-487b-47f5-86f8-04a6edc26ea7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.178047] env[62235]: DEBUG nova.compute.provider_tree [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.228763] env[62235]: DEBUG nova.network.neutron [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Updated VIF entry in instance network info cache for port 5f04dc27-a2d4-46cc-bf3c-947c4c3dab59. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.229200] env[62235]: DEBUG nova.network.neutron [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Updating instance_info_cache with network_info: [{"id": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "address": "fa:16:3e:9a:2c:29", "network": {"id": "c12598e4-11de-43b5-a569-b030b45411a7", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1353190335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "940941e982494a4caeebda900b3e0b8c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94926d5b-bfab-4c04-85b5-0fe89934c8ff", "external-id": "nsx-vlan-transportzone-157", "segmentation_id": 157, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f04dc27-a2", "ovs_interfaceid": "5f04dc27-a2d4-46cc-bf3c-947c4c3dab59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.274973] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271692, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.351396] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.351804] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.352074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.352297] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.352484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.355539] env[62235]: INFO nova.compute.manager [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Terminating instance [ 839.357894] env[62235]: DEBUG nova.compute.manager [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.358192] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.359221] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43dd411c-5d0d-4b40-bf73-df8b459f3ee4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.365616] env[62235]: DEBUG nova.compute.manager [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received event network-changed-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.366173] env[62235]: DEBUG nova.compute.manager [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing instance network info cache due to event network-changed-600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.366173] env[62235]: DEBUG oslo_concurrency.lockutils [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] Acquiring lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.372283] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.372717] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54f0102e-ffdc-4d24-98d9-2ff0f70e7316 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.381579] env[62235]: DEBUG oslo_vmware.api [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 839.381579] env[62235]: value = "task-1271694" [ 839.381579] env[62235]: _type = "Task" [ 839.381579] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.392710] env[62235]: DEBUG oslo_vmware.api [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.592917] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271693, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475413} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.593996] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9/9f5fcb83-bb6f-4815-b842-a37b2dbf26e9.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.594334] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.594719] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2391a402-81e0-44a8-92a9-b764bb0851cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.601787] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 839.601787] env[62235]: value = "task-1271695" [ 839.601787] env[62235]: _type = "Task" [ 839.601787] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.610514] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271695, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.681086] env[62235]: DEBUG nova.scheduler.client.report [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.734248] env[62235]: DEBUG oslo_concurrency.lockutils [req-461caba0-b6ef-48ed-8c71-f0d9fa0a720c req-ce3bc15a-dd13-480d-9fe0-30946e47204b service nova] Releasing lock "refresh_cache-80a803d1-c034-42b6-a15d-9fc3893953ad" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.750069] env[62235]: DEBUG nova.network.neutron [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updated VIF entry in instance network info cache for port 600609cd-3b8f-4683-9136-308a989872ec. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.750449] env[62235]: DEBUG nova.network.neutron [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [{"id": "600609cd-3b8f-4683-9136-308a989872ec", "address": "fa:16:3e:9f:b2:2e", "network": {"id": "e3892181-6713-45be-9572-f39a5a3955b2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-769806034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e340ac03d3734399a92abe4ed014251c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600609cd-3b", "ovs_interfaceid": "600609cd-3b8f-4683-9136-308a989872ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.777780] env[62235]: DEBUG oslo_vmware.api [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271692, 'name': PowerOnVM_Task, 'duration_secs': 0.631936} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.778028] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.778232] env[62235]: INFO nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Took 7.46 seconds to spawn the instance on the hypervisor. [ 839.778415] env[62235]: DEBUG nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.779265] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a25473-8186-4b82-804c-1cfcf7292af8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.891990] env[62235]: DEBUG oslo_vmware.api [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271694, 'name': PowerOffVM_Task, 'duration_secs': 0.314041} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.894915] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.894915] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.894915] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db3d4c57-3bc1-432c-a02c-58d6a10a6fcf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.962081] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.962081] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.962081] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Deleting the datastore file [datastore1] 876b7602-ccf0-4aed-90e2-d07fbf548fb6 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.962081] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a4c4e5e-70bb-4c8b-ac4f-28f55530c840 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.969800] env[62235]: DEBUG oslo_vmware.api [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for the task: (returnval){ [ 839.969800] env[62235]: value = "task-1271697" [ 839.969800] env[62235]: _type = "Task" [ 839.969800] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.977304] env[62235]: DEBUG oslo_vmware.api [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271697, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.031598] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Successfully updated port: eede8be1-1cbc-4791-802e-52193488eb27 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.114520] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271695, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070986} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.114848] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.115623] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e0ce6c-7e66-4d8e-b285-e902d116cc91 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.139689] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9/9f5fcb83-bb6f-4815-b842-a37b2dbf26e9.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.140009] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b61d0625-760d-470a-aeaa-09d48c86e737 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.160489] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 840.160489] env[62235]: value = "task-1271698" [ 840.160489] env[62235]: _type = "Task" [ 840.160489] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.168424] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271698, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.189667] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.190262] env[62235]: DEBUG nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.193782] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.340s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.198552] env[62235]: INFO nova.compute.claims [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.252841] env[62235]: DEBUG oslo_concurrency.lockutils [req-06c1e32e-e1af-47ac-9b28-1b33eaf989ab req-03841889-2491-4806-963d-b4edb0782bfa service nova] Releasing lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.253307] env[62235]: DEBUG oslo_concurrency.lockutils [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] Acquired lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.253499] env[62235]: DEBUG nova.network.neutron [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Refreshing network info cache for port 600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.297120] env[62235]: INFO nova.compute.manager [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Took 33.06 seconds to build instance. [ 840.421142] env[62235]: DEBUG nova.compute.manager [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received event network-vif-plugged-eede8be1-1cbc-4791-802e-52193488eb27 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.421597] env[62235]: DEBUG oslo_concurrency.lockutils [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] Acquiring lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.421597] env[62235]: DEBUG oslo_concurrency.lockutils [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.421839] env[62235]: DEBUG oslo_concurrency.lockutils [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.421930] env[62235]: DEBUG nova.compute.manager [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] No waiting events found dispatching network-vif-plugged-eede8be1-1cbc-4791-802e-52193488eb27 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.422111] env[62235]: WARNING nova.compute.manager [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received unexpected event network-vif-plugged-eede8be1-1cbc-4791-802e-52193488eb27 for instance with vm_state building and task_state spawning. [ 840.422283] env[62235]: DEBUG nova.compute.manager [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received event network-changed-eede8be1-1cbc-4791-802e-52193488eb27 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.422446] env[62235]: DEBUG nova.compute.manager [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Refreshing instance network info cache due to event network-changed-eede8be1-1cbc-4791-802e-52193488eb27. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.422634] env[62235]: DEBUG oslo_concurrency.lockutils [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] Acquiring lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.422774] env[62235]: DEBUG oslo_concurrency.lockutils [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] Acquired lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.422933] env[62235]: DEBUG nova.network.neutron [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Refreshing network info cache for port eede8be1-1cbc-4791-802e-52193488eb27 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.481570] env[62235]: DEBUG oslo_vmware.api [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Task: {'id': task-1271697, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196228} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.481730] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.481927] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.482308] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.482308] env[62235]: INFO nova.compute.manager [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 840.482536] env[62235]: DEBUG oslo.service.loopingcall [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.482739] env[62235]: DEBUG nova.compute.manager [-] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.482836] env[62235]: DEBUG nova.network.neutron [-] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.677372] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271698, 'name': ReconfigVM_Task, 'duration_secs': 0.290822} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.677510] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9/9f5fcb83-bb6f-4815-b842-a37b2dbf26e9.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.678118] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2faa3eb0-8ef9-4b15-8406-d5d670e71f80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.689631] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 840.689631] env[62235]: value = "task-1271699" [ 840.689631] env[62235]: _type = "Task" [ 840.689631] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.700355] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271699, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.705808] env[62235]: DEBUG nova.compute.utils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.712677] env[62235]: DEBUG nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.712677] env[62235]: DEBUG nova.network.neutron [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.781126] env[62235]: INFO nova.network.neutron [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Port 600609cd-3b8f-4683-9136-308a989872ec from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 840.781126] env[62235]: DEBUG nova.network.neutron [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.790130] env[62235]: DEBUG nova.policy [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96af7a114d04c0dad3c53f5a2dcd2bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef6547723ccf4be891da50a179dae06f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.799110] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1165b950-1ec3-463e-80a5-d1f0b71c2944 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.543s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.979350] env[62235]: DEBUG nova.network.neutron [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.085145] env[62235]: DEBUG nova.network.neutron [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.182399] env[62235]: DEBUG nova.network.neutron [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Successfully created port: 7bae7814-0344-4573-93e3-27f411ada206 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.201753] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271699, 'name': Rename_Task, 'duration_secs': 0.147294} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.202384] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.202917] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-593cb403-4c0e-4a38-ae76-76fae77ee844 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.211758] env[62235]: DEBUG nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.214433] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 841.214433] env[62235]: value = "task-1271700" [ 841.214433] env[62235]: _type = "Task" [ 841.214433] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.227118] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271700, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.248937] env[62235]: DEBUG nova.network.neutron [-] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.284742] env[62235]: DEBUG oslo_concurrency.lockutils [req-6fba9cfb-bd22-45f3-bd1c-5b1fe52e38f3 req-80897812-1b63-4fa7-954c-2b103745e9b6 service nova] Releasing lock "refresh_cache-876b7602-ccf0-4aed-90e2-d07fbf548fb6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.485267] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c2f3e6-3586-4809-b10f-9e182e6bd50a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.492634] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe668776-a9a7-450f-81c6-27007598e7eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.524521] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec6aa05-38e8-4410-9239-a3fa5500f8ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.532089] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc172405-5fc2-43bb-b93b-e9c4bdcf0a12 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.546635] env[62235]: DEBUG nova.compute.provider_tree [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.587440] env[62235]: DEBUG oslo_concurrency.lockutils [req-fce41f89-67ec-47ac-8b8a-1ebfe5b24ba7 req-4761b2da-fb3d-4ad1-b69e-c937ebadcc3a service nova] Releasing lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.729461] env[62235]: DEBUG oslo_vmware.api [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271700, 'name': PowerOnVM_Task, 'duration_secs': 0.485081} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.729819] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.730051] env[62235]: INFO nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Took 7.09 seconds to spawn the instance on the hypervisor. [ 841.730802] env[62235]: DEBUG nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.731743] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64e9b5a-12cc-4145-b981-84ee2e54ae71 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.753795] env[62235]: INFO nova.compute.manager [-] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Took 1.27 seconds to deallocate network for instance. [ 842.049953] env[62235]: DEBUG nova.scheduler.client.report [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.225428] env[62235]: DEBUG nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.253766] env[62235]: INFO nova.compute.manager [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Took 30.56 seconds to build instance. [ 842.258832] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.261359] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.261544] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.261748] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.261975] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.262120] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.262274] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.262491] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.262656] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.262834] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.262999] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.263191] env[62235]: DEBUG nova.virt.hardware [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.264129] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d38c0a-d647-456e-aa82-d39a3a7a036a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.267041] env[62235]: INFO nova.compute.manager [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Rebuilding instance [ 842.271583] env[62235]: DEBUG nova.compute.manager [req-dab94a62-c501-4368-a620-b225a3c633c5 req-49bd3eee-9a43-47f8-ae85-5fec4ce1aec3 service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received event network-vif-plugged-90607cc9-dfc7-4f25-b1dd-ad81a02691f9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.271583] env[62235]: DEBUG oslo_concurrency.lockutils [req-dab94a62-c501-4368-a620-b225a3c633c5 req-49bd3eee-9a43-47f8-ae85-5fec4ce1aec3 service nova] Acquiring lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.271583] env[62235]: DEBUG oslo_concurrency.lockutils [req-dab94a62-c501-4368-a620-b225a3c633c5 req-49bd3eee-9a43-47f8-ae85-5fec4ce1aec3 service nova] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.271889] env[62235]: DEBUG oslo_concurrency.lockutils [req-dab94a62-c501-4368-a620-b225a3c633c5 req-49bd3eee-9a43-47f8-ae85-5fec4ce1aec3 service nova] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.271889] env[62235]: DEBUG nova.compute.manager [req-dab94a62-c501-4368-a620-b225a3c633c5 req-49bd3eee-9a43-47f8-ae85-5fec4ce1aec3 service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] No waiting events found dispatching network-vif-plugged-90607cc9-dfc7-4f25-b1dd-ad81a02691f9 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.273863] env[62235]: WARNING nova.compute.manager [req-dab94a62-c501-4368-a620-b225a3c633c5 req-49bd3eee-9a43-47f8-ae85-5fec4ce1aec3 service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received unexpected event network-vif-plugged-90607cc9-dfc7-4f25-b1dd-ad81a02691f9 for instance with vm_state building and task_state spawning. [ 842.278643] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46b1f18-a7ba-47c3-ae1c-635e4fe5214f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.318921] env[62235]: DEBUG nova.compute.manager [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.319849] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be02e819-008e-4f6a-96e8-4e9a336b8f66 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.376507] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Successfully updated port: 90607cc9-dfc7-4f25-b1dd-ad81a02691f9 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.494253] env[62235]: DEBUG nova.compute.manager [req-2023edc3-c591-445b-88a8-a501f8c5eaa8 req-8efa6824-3a68-4a49-8a23-5fc5c294bb87 service nova] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Received event network-vif-deleted-600609cd-3b8f-4683-9136-308a989872ec {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.557034] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.557650] env[62235]: DEBUG nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.561214] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.277s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.562764] env[62235]: INFO nova.compute.claims [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.757701] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e6f8f53a-dbcc-4a2b-973c-07c44157d89a tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.134s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.799969] env[62235]: DEBUG nova.network.neutron [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Successfully updated port: 7bae7814-0344-4573-93e3-27f411ada206 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.834371] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.834913] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8c8849e-e75b-483d-bcd0-b000293f9280 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.843237] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 842.843237] env[62235]: value = "task-1271701" [ 842.843237] env[62235]: _type = "Task" [ 842.843237] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.851770] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271701, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.882624] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.882624] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquired lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.882747] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.069837] env[62235]: DEBUG nova.compute.utils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.071706] env[62235]: DEBUG nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.071706] env[62235]: DEBUG nova.network.neutron [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.146316] env[62235]: DEBUG nova.policy [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3ffdf6381df4aca8b4c613f13219262', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4456e079eae64f41b1596821d41bd275', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.305633] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-9474de53-1964-4a52-8783-2bac5773b6c9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.305633] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-9474de53-1964-4a52-8783-2bac5773b6c9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.305633] env[62235]: DEBUG nova.network.neutron [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.354205] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271701, 'name': PowerOffVM_Task, 'duration_secs': 0.351111} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.354646] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.355024] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 843.355858] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b1e94c-e3c5-433e-b915-4553f9b21889 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.362550] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.362888] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-717ab097-76d1-4f95-a8d0-a912a4d73ce1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.432716] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.432716] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.439977] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleting the datastore file [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.439977] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8c50358-473a-43bb-8f96-159e87b7e750 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.454568] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 843.454568] env[62235]: value = "task-1271703" [ 843.454568] env[62235]: _type = "Task" [ 843.454568] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.461112] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.480311] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.574608] env[62235]: DEBUG nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.717681] env[62235]: DEBUG nova.network.neutron [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Successfully created port: bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.846534] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c04e7a-7b62-468a-b46d-ace54c06df13 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.851133] env[62235]: DEBUG nova.network.neutron [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.859572] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a88251-4831-4240-98d7-04836b25472f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.902664] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d7546e-47a9-4378-b852-3ab38dd14eed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.918360] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2414418a-abdf-477b-b0e3-781aefa7085c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.931531] env[62235]: DEBUG nova.compute.provider_tree [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.971893] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161569} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.972217] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.972428] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.972612] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 844.108408] env[62235]: DEBUG nova.network.neutron [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Updating instance_info_cache with network_info: [{"id": "7bae7814-0344-4573-93e3-27f411ada206", "address": "fa:16:3e:8f:f5:b0", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bae7814-03", "ovs_interfaceid": "7bae7814-0344-4573-93e3-27f411ada206", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.160376] env[62235]: DEBUG nova.network.neutron [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Updating instance_info_cache with network_info: [{"id": "eede8be1-1cbc-4791-802e-52193488eb27", "address": "fa:16:3e:22:3e:77", "network": {"id": "67ccf44d-1843-430b-8f49-2a42543fb67e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1713431871", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.18", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91d149e97a7b4bdfa43ed32cb59d0fdf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeede8be1-1c", "ovs_interfaceid": "eede8be1-1cbc-4791-802e-52193488eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90607cc9-dfc7-4f25-b1dd-ad81a02691f9", "address": "fa:16:3e:30:4e:4f", "network": {"id": "7e4abc04-5d99-47d1-b8ef-246e27cdf50d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1984087248", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.124", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "91d149e97a7b4bdfa43ed32cb59d0fdf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90607cc9-df", "ovs_interfaceid": "90607cc9-dfc7-4f25-b1dd-ad81a02691f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.306366] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received event network-changed-90607cc9-dfc7-4f25-b1dd-ad81a02691f9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.306641] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Refreshing instance network info cache due to event network-changed-90607cc9-dfc7-4f25-b1dd-ad81a02691f9. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.307313] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Acquiring lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.332312] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.332650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.333228] env[62235]: INFO nova.compute.manager [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Rebooting instance [ 844.435065] env[62235]: DEBUG nova.scheduler.client.report [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.587441] env[62235]: DEBUG nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.612073] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.612337] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.612499] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.612688] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.612961] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.613191] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.613348] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.613673] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.613882] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.614066] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.614252] env[62235]: DEBUG nova.virt.hardware [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.614740] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-9474de53-1964-4a52-8783-2bac5773b6c9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.615032] env[62235]: DEBUG nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Instance network_info: |[{"id": "7bae7814-0344-4573-93e3-27f411ada206", "address": "fa:16:3e:8f:f5:b0", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bae7814-03", "ovs_interfaceid": "7bae7814-0344-4573-93e3-27f411ada206", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.615829] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b969f0-f9c7-4bd6-803c-dabc7e0b0a6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.618643] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:f5:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7bae7814-0344-4573-93e3-27f411ada206', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.625935] env[62235]: DEBUG oslo.service.loopingcall [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.626180] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.626906] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e341d2f-bf93-4358-9d5a-7307bbd554ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.645823] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6e59f8-0e38-4726-978e-f171950cfbd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.650532] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.650532] env[62235]: value = "task-1271704" [ 844.650532] env[62235]: _type = "Task" [ 844.650532] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.663570] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Releasing lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.663914] env[62235]: DEBUG nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Instance network_info: |[{"id": "eede8be1-1cbc-4791-802e-52193488eb27", "address": "fa:16:3e:22:3e:77", "network": {"id": "67ccf44d-1843-430b-8f49-2a42543fb67e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1713431871", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.18", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91d149e97a7b4bdfa43ed32cb59d0fdf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeede8be1-1c", "ovs_interfaceid": "eede8be1-1cbc-4791-802e-52193488eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90607cc9-dfc7-4f25-b1dd-ad81a02691f9", "address": "fa:16:3e:30:4e:4f", "network": {"id": "7e4abc04-5d99-47d1-b8ef-246e27cdf50d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1984087248", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.124", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "91d149e97a7b4bdfa43ed32cb59d0fdf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90607cc9-df", "ovs_interfaceid": "90607cc9-dfc7-4f25-b1dd-ad81a02691f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.666986] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Acquired lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.667186] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Refreshing network info cache for port 90607cc9-dfc7-4f25-b1dd-ad81a02691f9 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.668883] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:3e:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8059554c-499f-44b4-be06-29f80ec36b34', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eede8be1-1cbc-4791-802e-52193488eb27', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:4e:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90607cc9-dfc7-4f25-b1dd-ad81a02691f9', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.678762] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Creating folder: Project (91d149e97a7b4bdfa43ed32cb59d0fdf). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.678762] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271704, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.679332] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-223cce75-c20d-4b61-a72d-a7ffab9f53b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.690974] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Created folder: Project (91d149e97a7b4bdfa43ed32cb59d0fdf) in parent group-v273362. [ 844.691289] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Creating folder: Instances. Parent ref: group-v273443. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.691540] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a753be9-626a-405b-a426-9a12d84b9685 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.702327] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Created folder: Instances in parent group-v273443. [ 844.702593] env[62235]: DEBUG oslo.service.loopingcall [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.702829] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.703088] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-316531d1-d68a-4e08-9efc-271893f86681 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.729880] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.729880] env[62235]: value = "task-1271707" [ 844.729880] env[62235]: _type = "Task" [ 844.729880] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.738182] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271707, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.863323] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.863542] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquired lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.863730] env[62235]: DEBUG nova.network.neutron [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.940946] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.940946] env[62235]: DEBUG nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.943154] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.303s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.943389] env[62235]: DEBUG nova.objects.instance [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lazy-loading 'resources' on Instance uuid b5a1da09-2137-4434-a1a4-174b196b61b0 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.007617] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.007916] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.008144] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.008377] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.008558] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.008711] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.008966] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.009082] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.009262] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.009432] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.009611] env[62235]: DEBUG nova.virt.hardware [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.010552] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8be0a20-963f-44dd-afb2-059bad8e6721 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.020877] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcc002e-c763-45ac-a5ef-270e2e5a9184 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.035939] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:c1:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfd2c6bf-a775-49d3-8e77-b851443ec66b', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.042553] env[62235]: DEBUG oslo.service.loopingcall [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.042849] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.043084] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6709514-6c3f-44b1-9512-7e23bd25c601 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.064050] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.064050] env[62235]: value = "task-1271708" [ 845.064050] env[62235]: _type = "Task" [ 845.064050] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.072373] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271708, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.160536] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271704, 'name': CreateVM_Task, 'duration_secs': 0.336022} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.160769] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.161519] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.161787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.162009] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.162276] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45217110-384e-4ac7-bb8f-e7c12eb0b3e0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.167038] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 845.167038] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f29609-6231-965d-438f-64237706c130" [ 845.167038] env[62235]: _type = "Task" [ 845.167038] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.176780] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f29609-6231-965d-438f-64237706c130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.246419] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271707, 'name': CreateVM_Task, 'duration_secs': 0.512173} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.246666] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.247433] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.446225] env[62235]: DEBUG nova.compute.utils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.449940] env[62235]: DEBUG nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.450122] env[62235]: DEBUG nova.network.neutron [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.559284] env[62235]: DEBUG nova.policy [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8a1c9e076304498999fe4c6c1d89c69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a272872fcc55419fb474121b7fefd4c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.577019] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271708, 'name': CreateVM_Task, 'duration_secs': 0.355217} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.577019] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.577019] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.674694] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb095e8-0583-45f7-bdd5-d2429211cf9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.680915] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f29609-6231-965d-438f-64237706c130, 'name': SearchDatastore_Task, 'duration_secs': 0.010826} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.681605] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.681884] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.682146] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.682301] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.682490] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.682793] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.683163] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.683397] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8efe7ebd-a1fe-4fe3-b2ef-955da63b496d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.687075] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e93ccf86-5bb4-4216-aa11-9a78b43c1057 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.691860] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066d7a8a-59cc-4123-aa2a-1890fa0bea44 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.697523] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 845.697523] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ff61d0-5310-0e60-6b27-3d18350a0f41" [ 845.697523] env[62235]: _type = "Task" [ 845.697523] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.726474] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.726835] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.731272] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff43d162-2dad-4e77-88a1-94f7b85c126f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.734149] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d86d50-e1c7-4e54-b999-a61813462a39 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.746970] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ff61d0-5310-0e60-6b27-3d18350a0f41, 'name': SearchDatastore_Task, 'duration_secs': 0.009568} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.747321] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 845.747321] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520c219c-6439-35bc-c582-6981708f4a88" [ 845.747321] env[62235]: _type = "Task" [ 845.747321] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.747692] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.747988] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.748276] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.750059] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b48a754-13fe-4c73-8542-42f69d181d9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.754138] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.754573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.757776] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98ec1661-41a2-4651-913c-647da1e3e79b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.764915] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520c219c-6439-35bc-c582-6981708f4a88, 'name': SearchDatastore_Task, 'duration_secs': 0.011236} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.774234] env[62235]: DEBUG nova.compute.provider_tree [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.778134] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 845.778134] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525f2a4d-3427-c897-b2eb-9b9375db8995" [ 845.778134] env[62235]: _type = "Task" [ 845.778134] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.778134] env[62235]: DEBUG nova.network.neutron [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updating instance_info_cache with network_info: [{"id": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "address": "fa:16:3e:3c:21:f8", "network": {"id": "bcc80e56-2152-4dd0-829f-14cc48fd0e88", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1093381308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d1081d5b610424d833a5f7611c1e796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba39fe2-6d", "ovs_interfaceid": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.778134] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf3cd17f-6a44-4112-bce1-1f0dd419cbef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.783594] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Updated VIF entry in instance network info cache for port 90607cc9-dfc7-4f25-b1dd-ad81a02691f9. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.784259] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Updating instance_info_cache with network_info: [{"id": "eede8be1-1cbc-4791-802e-52193488eb27", "address": "fa:16:3e:22:3e:77", "network": {"id": "67ccf44d-1843-430b-8f49-2a42543fb67e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1713431871", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.18", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91d149e97a7b4bdfa43ed32cb59d0fdf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8059554c-499f-44b4-be06-29f80ec36b34", "external-id": "nsx-vlan-transportzone-892", "segmentation_id": 892, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeede8be1-1c", "ovs_interfaceid": "eede8be1-1cbc-4791-802e-52193488eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90607cc9-dfc7-4f25-b1dd-ad81a02691f9", "address": "fa:16:3e:30:4e:4f", "network": {"id": "7e4abc04-5d99-47d1-b8ef-246e27cdf50d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1984087248", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.124", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "91d149e97a7b4bdfa43ed32cb59d0fdf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90607cc9-df", "ovs_interfaceid": "90607cc9-dfc7-4f25-b1dd-ad81a02691f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.792144] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 845.792144] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52344c11-0ac7-040d-aa2c-42268b4fe804" [ 845.792144] env[62235]: _type = "Task" [ 845.792144] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.792985] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525f2a4d-3427-c897-b2eb-9b9375db8995, 'name': SearchDatastore_Task, 'duration_secs': 0.009181} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.793371] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.793602] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.793818] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.803517] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52344c11-0ac7-040d-aa2c-42268b4fe804, 'name': SearchDatastore_Task, 'duration_secs': 0.011696} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.804430] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.804430] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 9474de53-1964-4a52-8783-2bac5773b6c9/9474de53-1964-4a52-8783-2bac5773b6c9.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.804430] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.804615] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.804727] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5542b537-69f5-4446-9368-2b937e6bc96e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.807492] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6486264c-e8e5-4f3c-8297-afceedbd8fb3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.813280] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 845.813280] env[62235]: value = "task-1271709" [ 845.813280] env[62235]: _type = "Task" [ 845.813280] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.817336] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.817493] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.818505] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71b7f7cd-8e91-4d8f-abb5-976280e41ebd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.823767] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.827018] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 845.827018] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5200aec6-0b54-8254-9733-dca8458f8520" [ 845.827018] env[62235]: _type = "Task" [ 845.827018] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.834338] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5200aec6-0b54-8254-9733-dca8458f8520, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.907339] env[62235]: DEBUG nova.compute.manager [req-593592d1-ca36-4fc0-9810-2b11319176fc req-005955e4-6b80-408e-a8ee-b122402886ac service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Received event network-vif-plugged-bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.907601] env[62235]: DEBUG oslo_concurrency.lockutils [req-593592d1-ca36-4fc0-9810-2b11319176fc req-005955e4-6b80-408e-a8ee-b122402886ac service nova] Acquiring lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.908070] env[62235]: DEBUG oslo_concurrency.lockutils [req-593592d1-ca36-4fc0-9810-2b11319176fc req-005955e4-6b80-408e-a8ee-b122402886ac service nova] Lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.908070] env[62235]: DEBUG oslo_concurrency.lockutils [req-593592d1-ca36-4fc0-9810-2b11319176fc req-005955e4-6b80-408e-a8ee-b122402886ac service nova] Lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.908236] env[62235]: DEBUG nova.compute.manager [req-593592d1-ca36-4fc0-9810-2b11319176fc req-005955e4-6b80-408e-a8ee-b122402886ac service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] No waiting events found dispatching network-vif-plugged-bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 845.908542] env[62235]: WARNING nova.compute.manager [req-593592d1-ca36-4fc0-9810-2b11319176fc req-005955e4-6b80-408e-a8ee-b122402886ac service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Received unexpected event network-vif-plugged-bd95632a-e453-409f-ab45-023167eaf726 for instance with vm_state building and task_state spawning. [ 845.954977] env[62235]: DEBUG nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.054051] env[62235]: DEBUG nova.network.neutron [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Successfully created port: 9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.277981] env[62235]: DEBUG nova.scheduler.client.report [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.281802] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Releasing lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.284192] env[62235]: DEBUG nova.compute.manager [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.285163] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ff6f44-f9b3-43bd-9279-86765f956989 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.288873] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Releasing lock "refresh_cache-b126a2c9-62e9-49a1-91ae-7a703e5fa490" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.289544] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Received event network-vif-plugged-7bae7814-0344-4573-93e3-27f411ada206 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.289544] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Acquiring lock "9474de53-1964-4a52-8783-2bac5773b6c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.289732] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Lock "9474de53-1964-4a52-8783-2bac5773b6c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.289834] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Lock "9474de53-1964-4a52-8783-2bac5773b6c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.290030] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] No waiting events found dispatching network-vif-plugged-7bae7814-0344-4573-93e3-27f411ada206 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.290124] env[62235]: WARNING nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Received unexpected event network-vif-plugged-7bae7814-0344-4573-93e3-27f411ada206 for instance with vm_state building and task_state spawning. [ 846.290313] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Received event network-changed-7bae7814-0344-4573-93e3-27f411ada206 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.290444] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Refreshing instance network info cache due to event network-changed-7bae7814-0344-4573-93e3-27f411ada206. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.290633] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Acquiring lock "refresh_cache-9474de53-1964-4a52-8783-2bac5773b6c9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.290777] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Acquired lock "refresh_cache-9474de53-1964-4a52-8783-2bac5773b6c9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.292110] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Refreshing network info cache for port 7bae7814-0344-4573-93e3-27f411ada206 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.326998] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271709, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.339617] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5200aec6-0b54-8254-9733-dca8458f8520, 'name': SearchDatastore_Task, 'duration_secs': 0.008507} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.340721] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b89a39f-248e-4c60-b7e3-49ae5658f37a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.347661] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 846.347661] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526b38bf-5b75-f84f-9cd4-f7f2c9b85657" [ 846.347661] env[62235]: _type = "Task" [ 846.347661] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.348535] env[62235]: DEBUG nova.network.neutron [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Successfully updated port: bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.359946] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526b38bf-5b75-f84f-9cd4-f7f2c9b85657, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.452780] env[62235]: DEBUG nova.compute.manager [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Received event network-changed-bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.453225] env[62235]: DEBUG nova.compute.manager [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Refreshing instance network info cache due to event network-changed-bd95632a-e453-409f-ab45-023167eaf726. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.453483] env[62235]: DEBUG oslo_concurrency.lockutils [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] Acquiring lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.453483] env[62235]: DEBUG oslo_concurrency.lockutils [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] Acquired lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.453607] env[62235]: DEBUG nova.network.neutron [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Refreshing network info cache for port bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.785418] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.787806] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.824s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.789401] env[62235]: INFO nova.compute.claims [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.807408] env[62235]: INFO nova.scheduler.client.report [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Deleted allocations for instance b5a1da09-2137-4434-a1a4-174b196b61b0 [ 846.823374] env[62235]: DEBUG oslo_concurrency.lockutils [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.823374] env[62235]: DEBUG oslo_concurrency.lockutils [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.823456] env[62235]: DEBUG nova.compute.manager [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.823744] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578533} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.826596] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdf8c9b-9268-4e1e-a05a-2edcbf515a95 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.829413] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 9474de53-1964-4a52-8783-2bac5773b6c9/9474de53-1964-4a52-8783-2bac5773b6c9.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.829634] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.829895] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6121b21-ab98-4b2e-83ce-caa3b80390d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.836147] env[62235]: DEBUG nova.compute.manager [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62235) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 846.836722] env[62235]: DEBUG nova.objects.instance [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'flavor' on Instance uuid 3ea7db36-ebde-4523-b803-ad121a00269b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.839937] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 846.839937] env[62235]: value = "task-1271710" [ 846.839937] env[62235]: _type = "Task" [ 846.839937] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.848588] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271710, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.853477] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.860682] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526b38bf-5b75-f84f-9cd4-f7f2c9b85657, 'name': SearchDatastore_Task, 'duration_secs': 0.023068} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.860961] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.861222] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b126a2c9-62e9-49a1-91ae-7a703e5fa490/b126a2c9-62e9-49a1-91ae-7a703e5fa490.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.861517] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.861725] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.861953] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43bca390-a2ce-402a-a82f-d995fd5e4a15 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.863911] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2c8f7c5-d784-464b-9a72-fa1fc0892f69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.870835] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 846.870835] env[62235]: value = "task-1271711" [ 846.870835] env[62235]: _type = "Task" [ 846.870835] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.875840] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.876042] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.879593] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84f6f248-b909-4a6c-94a6-0b14203d8516 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.885809] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.888999] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 846.888999] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5201dde9-ed81-39ba-a984-33a48c7e766a" [ 846.888999] env[62235]: _type = "Task" [ 846.888999] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.897607] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5201dde9-ed81-39ba-a984-33a48c7e766a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.970184] env[62235]: DEBUG nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.006701] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.007162] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.007377] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.007578] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.007730] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.007889] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.008123] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.008296] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.008478] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.008651] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.008856] env[62235]: DEBUG nova.virt.hardware [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.009719] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de838085-2a77-4d5e-ac52-2fe0f35b76c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.018867] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418893a3-17e6-47c1-8100-7a5c6cb9b4da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.024126] env[62235]: DEBUG nova.network.neutron [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.224035] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Updated VIF entry in instance network info cache for port 7bae7814-0344-4573-93e3-27f411ada206. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.224035] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Updating instance_info_cache with network_info: [{"id": "7bae7814-0344-4573-93e3-27f411ada206", "address": "fa:16:3e:8f:f5:b0", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bae7814-03", "ovs_interfaceid": "7bae7814-0344-4573-93e3-27f411ada206", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.247394] env[62235]: DEBUG nova.network.neutron [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.309025] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2715aba-c2b6-4984-8563-4eecf3c7d337 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.316840] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c588c035-b295-410c-af10-240385273867 tempest-ServerDiagnosticsNegativeTest-1138975848 tempest-ServerDiagnosticsNegativeTest-1138975848-project-member] Lock "b5a1da09-2137-4434-a1a4-174b196b61b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.894s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.320309] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Doing hard reboot of VM {{(pid=62235) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 847.320859] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-f8d1c7d5-b89d-4e27-a93b-0642d77d7df9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.328841] env[62235]: DEBUG oslo_vmware.api [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 847.328841] env[62235]: value = "task-1271712" [ 847.328841] env[62235]: _type = "Task" [ 847.328841] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.338609] env[62235]: DEBUG oslo_vmware.api [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271712, 'name': ResetVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.346325] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.346562] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bcb7dfb3-3e9e-4cf0-a9a2-e3781471649c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.353568] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271710, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073223} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.354736] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.355080] env[62235]: DEBUG oslo_vmware.api [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 847.355080] env[62235]: value = "task-1271713" [ 847.355080] env[62235]: _type = "Task" [ 847.355080] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.355747] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48992c72-fd5d-4d1a-ba8a-00a16c10c70d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.365773] env[62235]: DEBUG oslo_vmware.api [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.384479] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 9474de53-1964-4a52-8783-2bac5773b6c9/9474de53-1964-4a52-8783-2bac5773b6c9.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.389723] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e63f3e25-b973-453c-a1c2-4c796c9a02a1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.414011] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271711, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465953} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.418505] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b126a2c9-62e9-49a1-91ae-7a703e5fa490/b126a2c9-62e9-49a1-91ae-7a703e5fa490.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.418744] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.419036] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5201dde9-ed81-39ba-a984-33a48c7e766a, 'name': SearchDatastore_Task, 'duration_secs': 0.009383} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.419302] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 847.419302] env[62235]: value = "task-1271714" [ 847.419302] env[62235]: _type = "Task" [ 847.419302] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.419493] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-785e885a-5cfd-4f67-a8d0-f02a24aed42d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.421965] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d70d62f-1070-4e7f-850a-308eeea4c11e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.430424] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 847.430424] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f01b0f-ba19-c6e7-70fd-35b2c7b163d0" [ 847.430424] env[62235]: _type = "Task" [ 847.430424] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.433765] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271714, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.437944] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 847.437944] env[62235]: value = "task-1271715" [ 847.437944] env[62235]: _type = "Task" [ 847.437944] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.445744] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f01b0f-ba19-c6e7-70fd-35b2c7b163d0, 'name': SearchDatastore_Task, 'duration_secs': 0.011031} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.446459] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.446734] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.447075] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d3f9a44-6450-4a03-8e07-d35a7d186c78 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.452011] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271715, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.456367] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 847.456367] env[62235]: value = "task-1271716" [ 847.456367] env[62235]: _type = "Task" [ 847.456367] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.464334] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271716, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.674660] env[62235]: DEBUG nova.compute.manager [req-5243ab3d-ca92-45ee-8685-e268f2d713cd req-e3aecaed-171f-4b6a-a9b4-03c80bbb28c6 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Received event network-vif-plugged-9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.675381] env[62235]: DEBUG oslo_concurrency.lockutils [req-5243ab3d-ca92-45ee-8685-e268f2d713cd req-e3aecaed-171f-4b6a-a9b4-03c80bbb28c6 service nova] Acquiring lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.675607] env[62235]: DEBUG oslo_concurrency.lockutils [req-5243ab3d-ca92-45ee-8685-e268f2d713cd req-e3aecaed-171f-4b6a-a9b4-03c80bbb28c6 service nova] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.675815] env[62235]: DEBUG oslo_concurrency.lockutils [req-5243ab3d-ca92-45ee-8685-e268f2d713cd req-e3aecaed-171f-4b6a-a9b4-03c80bbb28c6 service nova] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.676011] env[62235]: DEBUG nova.compute.manager [req-5243ab3d-ca92-45ee-8685-e268f2d713cd req-e3aecaed-171f-4b6a-a9b4-03c80bbb28c6 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] No waiting events found dispatching network-vif-plugged-9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.676195] env[62235]: WARNING nova.compute.manager [req-5243ab3d-ca92-45ee-8685-e268f2d713cd req-e3aecaed-171f-4b6a-a9b4-03c80bbb28c6 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Received unexpected event network-vif-plugged-9af619cc-4e60-4097-96e6-b957861ba812 for instance with vm_state building and task_state spawning. [ 847.727573] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Releasing lock "refresh_cache-9474de53-1964-4a52-8783-2bac5773b6c9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.727890] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Received event network-changed-5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.728022] env[62235]: DEBUG nova.compute.manager [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Refreshing instance network info cache due to event network-changed-5ba39fe2-6d0a-4a24-8b3f-024002027e22. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.728353] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Acquiring lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.728527] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Acquired lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.728707] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Refreshing network info cache for port 5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.750135] env[62235]: DEBUG oslo_concurrency.lockutils [req-a3a5b5e1-44b1-4ee0-9753-d8b5cd52bb4f req-5a68bc88-5f1b-44be-8fc5-943575e265ae service nova] Releasing lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.750480] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.750644] env[62235]: DEBUG nova.network.neutron [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.772513] env[62235]: DEBUG nova.network.neutron [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Successfully updated port: 9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.842310] env[62235]: DEBUG oslo_vmware.api [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271712, 'name': ResetVM_Task, 'duration_secs': 0.102922} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.846283] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Did hard reboot of VM {{(pid=62235) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 847.846577] env[62235]: DEBUG nova.compute.manager [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.847816] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c6d3ab-1479-4756-97d1-f40ee1afc92c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.870454] env[62235]: DEBUG oslo_vmware.api [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271713, 'name': PowerOffVM_Task, 'duration_secs': 0.266941} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.870847] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.870942] env[62235]: DEBUG nova.compute.manager [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.872205] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7196ea9-42d2-40ea-9978-43dfb432ca37 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.937497] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271714, 'name': ReconfigVM_Task, 'duration_secs': 0.470025} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.937676] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 9474de53-1964-4a52-8783-2bac5773b6c9/9474de53-1964-4a52-8783-2bac5773b6c9.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.938513] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef05e0c2-0d30-4b8a-9de0-430fbb8952f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.952978] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271715, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073589} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.954421] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.954837] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 847.954837] env[62235]: value = "task-1271717" [ 847.954837] env[62235]: _type = "Task" [ 847.954837] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.955573] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db81e7d-5894-49aa-a154-3b25dec70bec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.993512] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] b126a2c9-62e9-49a1-91ae-7a703e5fa490/b126a2c9-62e9-49a1-91ae-7a703e5fa490.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.003238] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6da23ce-0ad6-4c73-b3cc-6962424ff323 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.017615] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271717, 'name': Rename_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.017892] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271716, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501712} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.018701] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.018932] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.019620] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7c1b0de-4178-4495-92f7-e6f63b4b79e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.024916] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 848.024916] env[62235]: value = "task-1271718" [ 848.024916] env[62235]: _type = "Task" [ 848.024916] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.026218] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 848.026218] env[62235]: value = "task-1271719" [ 848.026218] env[62235]: _type = "Task" [ 848.026218] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.041571] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271718, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.044855] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271719, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.108783] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aada272-b38d-418f-9972-2c17b66cbb45 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.116143] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81326b65-a01c-4fef-827f-987ac2a4b975 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.145877] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d75b1d4-7e59-4a68-9e6b-73b648e03aa6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.153360] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a3e615-a7d9-4dc1-9f9a-8aba60ad48cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.166214] env[62235]: DEBUG nova.compute.provider_tree [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.274437] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.274618] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.274841] env[62235]: DEBUG nova.network.neutron [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.303599] env[62235]: DEBUG nova.network.neutron [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.361728] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8564268c-aced-4ff7-831d-463b9f04f16f tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.028s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.388984] env[62235]: DEBUG oslo_concurrency.lockutils [None req-04654cb0-fa7a-4dc7-a9e6-9f2a76e3d3d7 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.566s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.471251] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271717, 'name': Rename_Task, 'duration_secs': 0.197711} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.471588] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.471843] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ad54ebf-4029-4c44-a1fb-e0cc3db4fb29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.479966] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 848.479966] env[62235]: value = "task-1271720" [ 848.479966] env[62235]: _type = "Task" [ 848.479966] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.489375] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271720, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.508276] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updated VIF entry in instance network info cache for port 5ba39fe2-6d0a-4a24-8b3f-024002027e22. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 848.508645] env[62235]: DEBUG nova.network.neutron [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updating instance_info_cache with network_info: [{"id": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "address": "fa:16:3e:3c:21:f8", "network": {"id": "bcc80e56-2152-4dd0-829f-14cc48fd0e88", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1093381308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d1081d5b610424d833a5f7611c1e796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba39fe2-6d", "ovs_interfaceid": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.536197] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271718, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.541046] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271719, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072227} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.541325] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.542088] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd541a97-d048-45cc-9314-eedb9ae802ec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.563894] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.564871] env[62235]: DEBUG nova.network.neutron [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updating instance_info_cache with network_info: [{"id": "bd95632a-e453-409f-ab45-023167eaf726", "address": "fa:16:3e:ef:ab:a1", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd95632a-e4", "ovs_interfaceid": "bd95632a-e453-409f-ab45-023167eaf726", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.567589] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e347732c-1579-4e19-8953-d8f74e36dacc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.581601] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Releasing lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.581893] env[62235]: DEBUG nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance network_info: |[{"id": "bd95632a-e453-409f-ab45-023167eaf726", "address": "fa:16:3e:ef:ab:a1", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd95632a-e4", "ovs_interfaceid": "bd95632a-e453-409f-ab45-023167eaf726", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.582332] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:ab:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '88651df2-0506-4f6c-b868-dd30a81f2b1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd95632a-e453-409f-ab45-023167eaf726', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.589650] env[62235]: DEBUG oslo.service.loopingcall [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.590574] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.590813] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-105897de-52b5-4641-876b-a406dc9efe9f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.607288] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 848.607288] env[62235]: value = "task-1271721" [ 848.607288] env[62235]: _type = "Task" [ 848.607288] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.614948] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.614948] env[62235]: value = "task-1271722" [ 848.614948] env[62235]: _type = "Task" [ 848.614948] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.619099] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271721, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.624377] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271722, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.669029] env[62235]: DEBUG nova.scheduler.client.report [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.825829] env[62235]: DEBUG nova.network.neutron [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.994934] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271720, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.012301] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc96dec6-e522-4fb3-abc9-3ebc56db2aea req-38abf72a-f565-4e04-8352-38e3d6513e9f service nova] Releasing lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.034055] env[62235]: DEBUG nova.network.neutron [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updating instance_info_cache with network_info: [{"id": "9af619cc-4e60-4097-96e6-b957861ba812", "address": "fa:16:3e:be:96:6b", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9af619cc-4e", "ovs_interfaceid": "9af619cc-4e60-4097-96e6-b957861ba812", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.039413] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271718, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.121793] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271721, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.129039] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271722, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.174300] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.174942] env[62235]: DEBUG nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.179075] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.797s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.179330] env[62235]: DEBUG nova.objects.instance [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lazy-loading 'resources' on Instance uuid ec1830c7-7095-45d1-a188-c2993c60035e {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.248532] env[62235]: DEBUG nova.objects.instance [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'flavor' on Instance uuid 3ea7db36-ebde-4523-b803-ad121a00269b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.493468] env[62235]: DEBUG oslo_vmware.api [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271720, 'name': PowerOnVM_Task, 'duration_secs': 0.864869} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.493951] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.493951] env[62235]: INFO nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Took 7.27 seconds to spawn the instance on the hypervisor. [ 849.494143] env[62235]: DEBUG nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.495040] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f111929b-e4f6-4b06-872b-15816c457548 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.538414] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271718, 'name': ReconfigVM_Task, 'duration_secs': 1.025178} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.539892] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Reconfigured VM instance instance-00000046 to attach disk [datastore1] b126a2c9-62e9-49a1-91ae-7a703e5fa490/b126a2c9-62e9-49a1-91ae-7a703e5fa490.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.540525] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.540805] env[62235]: DEBUG nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Instance network_info: |[{"id": "9af619cc-4e60-4097-96e6-b957861ba812", "address": "fa:16:3e:be:96:6b", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9af619cc-4e", "ovs_interfaceid": "9af619cc-4e60-4097-96e6-b957861ba812", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.541370] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3ff351e-fe4c-49df-8db9-5716af8aaeee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.543724] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:96:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9af619cc-4e60-4097-96e6-b957861ba812', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.551767] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Creating folder: Project (a272872fcc55419fb474121b7fefd4c5). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.552530] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30661170-ba40-409c-9a9e-ff7904a970aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.559938] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 849.559938] env[62235]: value = "task-1271723" [ 849.559938] env[62235]: _type = "Task" [ 849.559938] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.565991] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Created folder: Project (a272872fcc55419fb474121b7fefd4c5) in parent group-v273362. [ 849.566222] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Creating folder: Instances. Parent ref: group-v273448. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.566781] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e436c300-bc38-43e4-8d2e-f2a952fdec79 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.572148] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271723, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.580080] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Created folder: Instances in parent group-v273448. [ 849.580080] env[62235]: DEBUG oslo.service.loopingcall [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.580080] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.580080] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-778322df-2771-44df-af9d-f967373b89ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.601883] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.601883] env[62235]: value = "task-1271726" [ 849.601883] env[62235]: _type = "Task" [ 849.601883] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.610029] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271726, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.619752] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271721, 'name': ReconfigVM_Task, 'duration_secs': 0.605133} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.623043] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c/00825c9e-b398-4edc-a5ed-e4b5b476af1c.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.623043] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd7c6870-a9b0-4b69-8333-f93ffe374bb4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.625552] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271722, 'name': CreateVM_Task, 'duration_secs': 0.584367} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.625987] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.626644] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.626799] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.627153] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.627400] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7b504c4-1c9e-4401-bbba-9b1c29305f5a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.630407] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 849.630407] env[62235]: value = "task-1271727" [ 849.630407] env[62235]: _type = "Task" [ 849.630407] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.637044] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 849.637044] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dde8a7-49a4-86cc-0b16-e8f04d832b16" [ 849.637044] env[62235]: _type = "Task" [ 849.637044] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.643525] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271727, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.648036] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dde8a7-49a4-86cc-0b16-e8f04d832b16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.682627] env[62235]: DEBUG nova.compute.utils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.687731] env[62235]: DEBUG nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.687845] env[62235]: DEBUG nova.network.neutron [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 849.753595] env[62235]: DEBUG oslo_concurrency.lockutils [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.753780] env[62235]: DEBUG oslo_concurrency.lockutils [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.753959] env[62235]: DEBUG nova.network.neutron [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.754171] env[62235]: DEBUG nova.objects.instance [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'info_cache' on Instance uuid 3ea7db36-ebde-4523-b803-ad121a00269b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.761761] env[62235]: DEBUG nova.compute.manager [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Received event network-changed-9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.762078] env[62235]: DEBUG nova.compute.manager [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Refreshing instance network info cache due to event network-changed-9af619cc-4e60-4097-96e6-b957861ba812. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.762351] env[62235]: DEBUG oslo_concurrency.lockutils [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] Acquiring lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.762552] env[62235]: DEBUG oslo_concurrency.lockutils [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] Acquired lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.762786] env[62235]: DEBUG nova.network.neutron [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Refreshing network info cache for port 9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.803631] env[62235]: DEBUG nova.policy [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d7ef1bc26534632ad28710f9df9803c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e22686dbba4420dad98d5de1f4fd449', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.910719] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.911101] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.911375] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.911681] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.911875] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.914422] env[62235]: INFO nova.compute.manager [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Terminating instance [ 849.916467] env[62235]: DEBUG nova.compute.manager [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.916730] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 849.918293] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35aa015-ff78-47bb-b95f-6da7c568be8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.931254] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 849.932037] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e55011e5-1808-49ac-aaa6-a8b2308c3987 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.938624] env[62235]: DEBUG oslo_vmware.api [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 849.938624] env[62235]: value = "task-1271728" [ 849.938624] env[62235]: _type = "Task" [ 849.938624] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.949731] env[62235]: DEBUG oslo_vmware.api [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.972651] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b78308-fd1f-4e96-a6cb-03c45cee48f9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.981270] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f813dd5-1e2d-47b1-9a4e-1d86aa1a7a52 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.019568] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e474a721-1700-4c56-a3f1-7ec514ed248b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.022157] env[62235]: INFO nova.compute.manager [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Took 26.55 seconds to build instance. [ 850.029049] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e953bd3a-f2ce-4133-be38-a8900311f246 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.046952] env[62235]: DEBUG nova.compute.provider_tree [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.069448] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271723, 'name': Rename_Task, 'duration_secs': 0.225182} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.069735] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.069987] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a525e739-efa6-4ff0-935e-38c50e7c2306 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.076112] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 850.076112] env[62235]: value = "task-1271729" [ 850.076112] env[62235]: _type = "Task" [ 850.076112] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.083902] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.114191] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271726, 'name': CreateVM_Task, 'duration_secs': 0.46163} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.114528] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.115277] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.115277] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.115580] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.115830] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f3fb1eb-f874-454e-8f1c-9583f5bcd8ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.122077] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 850.122077] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a33c98-8e8a-a412-ec03-bde27bc3909f" [ 850.122077] env[62235]: _type = "Task" [ 850.122077] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.131432] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a33c98-8e8a-a412-ec03-bde27bc3909f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.139960] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271727, 'name': Rename_Task, 'duration_secs': 0.223229} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.143184] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.143383] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e5d8074-66b4-4f4d-a6b8-c2e544ea81f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.153447] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dde8a7-49a4-86cc-0b16-e8f04d832b16, 'name': SearchDatastore_Task, 'duration_secs': 0.026064} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.156097] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.156097] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.156097] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.156097] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.156097] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.156097] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 850.156097] env[62235]: value = "task-1271730" [ 850.156097] env[62235]: _type = "Task" [ 850.156097] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.156598] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2202a81c-5a83-45cb-87c7-51869522886d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.166303] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271730, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.169051] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.169051] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.169976] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06ed78b5-a62c-414b-8bd5-94a21c94a7a0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.175451] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 850.175451] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52679cbe-e46c-9544-88c0-33bb32270e75" [ 850.175451] env[62235]: _type = "Task" [ 850.175451] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.184291] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52679cbe-e46c-9544-88c0-33bb32270e75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.187988] env[62235]: DEBUG nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.261583] env[62235]: DEBUG nova.objects.base [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Object Instance<3ea7db36-ebde-4523-b803-ad121a00269b> lazy-loaded attributes: flavor,info_cache {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 850.403182] env[62235]: DEBUG nova.network.neutron [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Successfully created port: 4c773abd-ae23-4075-aa4d-efb7b61a598c {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.453097] env[62235]: DEBUG oslo_vmware.api [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271728, 'name': PowerOffVM_Task, 'duration_secs': 0.324233} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.453440] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.453622] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.453879] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d15e6cb-7a22-4ac2-915d-b87fbfc0b658 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.484445] env[62235]: DEBUG nova.network.neutron [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updated VIF entry in instance network info cache for port 9af619cc-4e60-4097-96e6-b957861ba812. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.484905] env[62235]: DEBUG nova.network.neutron [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updating instance_info_cache with network_info: [{"id": "9af619cc-4e60-4097-96e6-b957861ba812", "address": "fa:16:3e:be:96:6b", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9af619cc-4e", "ovs_interfaceid": "9af619cc-4e60-4097-96e6-b957861ba812", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.513159] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 850.513522] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 850.513584] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Deleting the datastore file [datastore1] 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.513813] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-140f5d9f-da47-40d0-ae71-78cede5bff0c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.520617] env[62235]: DEBUG oslo_vmware.api [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for the task: (returnval){ [ 850.520617] env[62235]: value = "task-1271732" [ 850.520617] env[62235]: _type = "Task" [ 850.520617] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.528672] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4de29cb8-d6b7-4ae4-9036-115a70dc3673 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.077s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.528978] env[62235]: DEBUG oslo_vmware.api [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271732, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.550488] env[62235]: DEBUG nova.scheduler.client.report [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.588767] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271729, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.633083] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a33c98-8e8a-a412-ec03-bde27bc3909f, 'name': SearchDatastore_Task, 'duration_secs': 0.009887} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.633403] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.633687] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.634150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.634319] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.634529] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.635192] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0534657a-6d2b-41ac-82dc-f669c3228967 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.643490] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.643686] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.644460] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76ffe861-c728-4314-9de4-3901a0182920 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.650599] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 850.650599] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b6e4e1-2f20-2891-75ac-92ac45989c63" [ 850.650599] env[62235]: _type = "Task" [ 850.650599] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.660025] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b6e4e1-2f20-2891-75ac-92ac45989c63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.669724] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271730, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.684464] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52679cbe-e46c-9544-88c0-33bb32270e75, 'name': SearchDatastore_Task, 'duration_secs': 0.009726} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.685310] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fb990a7-0df1-4468-9452-0b37ebdcbbd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.690965] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 850.690965] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523bc067-0c3d-8533-a06c-257e9e8dcbb0" [ 850.690965] env[62235]: _type = "Task" [ 850.690965] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.702736] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523bc067-0c3d-8533-a06c-257e9e8dcbb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.987282] env[62235]: DEBUG oslo_concurrency.lockutils [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] Releasing lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.987524] env[62235]: DEBUG nova.compute.manager [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Received event network-changed-5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.987725] env[62235]: DEBUG nova.compute.manager [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Refreshing instance network info cache due to event network-changed-5ba39fe2-6d0a-4a24-8b3f-024002027e22. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.988281] env[62235]: DEBUG oslo_concurrency.lockutils [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] Acquiring lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.988281] env[62235]: DEBUG oslo_concurrency.lockutils [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] Acquired lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.988380] env[62235]: DEBUG nova.network.neutron [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Refreshing network info cache for port 5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.045041] env[62235]: DEBUG oslo_vmware.api [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Task: {'id': task-1271732, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137442} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.045041] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.045041] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.045275] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.045313] env[62235]: INFO nova.compute.manager [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 851.045791] env[62235]: DEBUG oslo.service.loopingcall [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.045791] env[62235]: DEBUG nova.compute.manager [-] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.046756] env[62235]: DEBUG nova.network.neutron [-] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.057493] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.060675] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.801s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.060675] env[62235]: DEBUG nova.objects.instance [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lazy-loading 'resources' on Instance uuid 876b7602-ccf0-4aed-90e2-d07fbf548fb6 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.083169] env[62235]: DEBUG nova.network.neutron [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.091170] env[62235]: DEBUG oslo_vmware.api [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271729, 'name': PowerOnVM_Task, 'duration_secs': 0.634371} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.092264] env[62235]: INFO nova.scheduler.client.report [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Deleted allocations for instance ec1830c7-7095-45d1-a188-c2993c60035e [ 851.093446] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.093666] env[62235]: INFO nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Took 13.11 seconds to spawn the instance on the hypervisor. [ 851.093843] env[62235]: DEBUG nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.096804] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19705c50-c444-4d63-891b-94317fc8a37d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.164395] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b6e4e1-2f20-2891-75ac-92ac45989c63, 'name': SearchDatastore_Task, 'duration_secs': 0.014483} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.169023] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-935e0493-3623-45e7-a36d-ab5736334765 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.173817] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "9474de53-1964-4a52-8783-2bac5773b6c9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.174147] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.181955] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271730, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.182168] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 851.182168] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b5be80-2027-fcff-dad6-4b6ee48e8655" [ 851.182168] env[62235]: _type = "Task" [ 851.182168] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.192399] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b5be80-2027-fcff-dad6-4b6ee48e8655, 'name': SearchDatastore_Task, 'duration_secs': 0.0101} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.195421] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.195775] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 15d41c0f-045d-4e1c-88b5-1841f99f2e74/15d41c0f-045d-4e1c-88b5-1841f99f2e74.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.196571] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c4ef500-8c51-47a8-8aa1-7c29e1ce7c6d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.200255] env[62235]: DEBUG nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.207425] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523bc067-0c3d-8533-a06c-257e9e8dcbb0, 'name': SearchDatastore_Task, 'duration_secs': 0.011508} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.207746] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.208103] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b7801744-5552-49b8-a95d-41b67bdf2380/b7801744-5552-49b8-a95d-41b67bdf2380.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.208408] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6f0316d-23e3-4297-9a88-692ccf185199 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.214430] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 851.214430] env[62235]: value = "task-1271733" [ 851.214430] env[62235]: _type = "Task" [ 851.214430] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.219801] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 851.219801] env[62235]: value = "task-1271734" [ 851.219801] env[62235]: _type = "Task" [ 851.219801] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.227582] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.236079] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271734, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.240502] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='9d49489d92a19c816c293cde5011ceac',container_format='bare',created_at=2024-10-12T06:23:16Z,direct_url=,disk_format='vmdk',id=4b23c73d-e39b-4f1a-8817-627621a5ee39,min_disk=1,min_ram=0,name='tempest-test-snap-243589479',owner='1e22686dbba4420dad98d5de1f4fd449',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-12T06:23:31Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.240869] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.241356] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.241492] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.241757] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.242083] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.242357] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.242616] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.243012] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.243139] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.243447] env[62235]: DEBUG nova.virt.hardware [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.244691] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b487a3-5169-4069-8d40-3b54b1301dd4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.253574] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35ed5cc-af88-4eb8-9340-8bd58013b82b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.587719] env[62235]: DEBUG oslo_concurrency.lockutils [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.605683] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b137e1c1-2a58-463f-ba56-f64166b94561 tempest-ServerShowV247Test-666894622 tempest-ServerShowV247Test-666894622-project-member] Lock "ec1830c7-7095-45d1-a188-c2993c60035e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.974s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.614806] env[62235]: INFO nova.compute.manager [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Took 35.82 seconds to build instance. [ 851.684021] env[62235]: DEBUG oslo_vmware.api [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271730, 'name': PowerOnVM_Task, 'duration_secs': 1.124996} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.685666] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.686546] env[62235]: DEBUG nova.compute.manager [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.687881] env[62235]: DEBUG nova.compute.utils [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.692117] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e7bb7c-0308-4708-a5d5-f4591d0e8c74 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.736239] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271734, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.736541] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271733, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.892478] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a05831e-7535-4c4c-968e-d603dac5e8dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.903611] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87afb794-fe3f-4241-b0a2-49831488eb19 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.948251] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38e4fea-606c-4f6f-a612-588c9722ffe6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.956802] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ddff2a-02b6-4991-b7de-4ecc985a2e8e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.971966] env[62235]: DEBUG nova.compute.provider_tree [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.994016] env[62235]: DEBUG nova.network.neutron [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updated VIF entry in instance network info cache for port 5ba39fe2-6d0a-4a24-8b3f-024002027e22. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.994412] env[62235]: DEBUG nova.network.neutron [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updating instance_info_cache with network_info: [{"id": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "address": "fa:16:3e:3c:21:f8", "network": {"id": "bcc80e56-2152-4dd0-829f-14cc48fd0e88", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1093381308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d1081d5b610424d833a5f7611c1e796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba39fe2-6d", "ovs_interfaceid": "5ba39fe2-6d0a-4a24-8b3f-024002027e22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.092512] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.092829] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f22c133-cccd-491e-bfd9-eac30837232b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.100156] env[62235]: DEBUG oslo_vmware.api [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 852.100156] env[62235]: value = "task-1271735" [ 852.100156] env[62235]: _type = "Task" [ 852.100156] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.108657] env[62235]: DEBUG oslo_vmware.api [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271735, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.117640] env[62235]: DEBUG oslo_concurrency.lockutils [None req-89486e3d-98e2-40c5-8565-0214b0e6967f tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.176s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.197111] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.023s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.212781] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.225541] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271733, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581697} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.229459] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 15d41c0f-045d-4e1c-88b5-1841f99f2e74/15d41c0f-045d-4e1c-88b5-1841f99f2e74.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.229771] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.230050] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c2a32cd-bb3d-4205-91da-466b933ceeb3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.239501] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271734, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.583012} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.240985] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b7801744-5552-49b8-a95d-41b67bdf2380/b7801744-5552-49b8-a95d-41b67bdf2380.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.241280] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.241695] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 852.241695] env[62235]: value = "task-1271736" [ 852.241695] env[62235]: _type = "Task" [ 852.241695] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.241904] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5bd2cf3-bf4c-4f97-9653-ffb8da94fde8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.256359] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271736, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.258263] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 852.258263] env[62235]: value = "task-1271737" [ 852.258263] env[62235]: _type = "Task" [ 852.258263] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.266355] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271737, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.360632] env[62235]: DEBUG nova.network.neutron [-] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.475569] env[62235]: DEBUG nova.scheduler.client.report [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.497232] env[62235]: DEBUG oslo_concurrency.lockutils [req-5cd594d1-0138-44a8-9147-1d82a91acf79 req-c22957fc-cd98-4a41-8d31-41012951039f service nova] Releasing lock "refresh_cache-9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.612886] env[62235]: DEBUG oslo_vmware.api [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271735, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.714829] env[62235]: DEBUG nova.compute.manager [req-056089f2-3a57-4ad7-9143-fb4c0fa41cf5 req-53deb8cb-9f12-47c6-a5d6-e6b4acf3b26d service nova] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Received event network-vif-deleted-5ba39fe2-6d0a-4a24-8b3f-024002027e22 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.745178] env[62235]: DEBUG nova.network.neutron [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Successfully updated port: 4c773abd-ae23-4075-aa4d-efb7b61a598c {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.759037] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271736, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.206403} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.762535] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.764031] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953d9d5e-9a01-47e9-953e-589add6bae68 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.775178] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271737, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.223052} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.787485] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.802659] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 15d41c0f-045d-4e1c-88b5-1841f99f2e74/15d41c0f-045d-4e1c-88b5-1841f99f2e74.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.803237] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b70e80b-b60c-431c-8587-a5a21a53ebe0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.809862] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d32e80b8-abf7-4361-816f-2d2d16ebdb70 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.848209] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] b7801744-5552-49b8-a95d-41b67bdf2380/b7801744-5552-49b8-a95d-41b67bdf2380.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.850457] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b59f8ef-8450-49f9-97c3-2e4a2ab9e82b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.867194] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 852.867194] env[62235]: value = "task-1271738" [ 852.867194] env[62235]: _type = "Task" [ 852.867194] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.867935] env[62235]: INFO nova.compute.manager [-] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Took 1.82 seconds to deallocate network for instance. [ 852.880277] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 852.880277] env[62235]: value = "task-1271739" [ 852.880277] env[62235]: _type = "Task" [ 852.880277] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.892453] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.900122] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.982404] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.922s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.986489] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.774s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.986679] env[62235]: DEBUG nova.objects.instance [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 853.020850] env[62235]: INFO nova.scheduler.client.report [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Deleted allocations for instance 876b7602-ccf0-4aed-90e2-d07fbf548fb6 [ 853.111871] env[62235]: DEBUG oslo_vmware.api [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271735, 'name': PowerOnVM_Task, 'duration_secs': 0.801553} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.112179] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.112370] env[62235]: DEBUG nova.compute.manager [None req-af02c4db-210d-4e6d-8546-c8827fb7787b tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.113159] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73ad2dc-be86-453e-91b3-449cb1cffc3c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.255074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "refresh_cache-c571a362-9b55-42d3-a019-1695c1caf248" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.255074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "refresh_cache-c571a362-9b55-42d3-a019-1695c1caf248" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.255074] env[62235]: DEBUG nova.network.neutron [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.327823] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "9474de53-1964-4a52-8783-2bac5773b6c9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.328182] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.328852] env[62235]: INFO nova.compute.manager [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Attaching volume 083d43a4-743a-4f4c-872c-16b184c0c86c to /dev/sdb [ 853.376402] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.388339] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.394519] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5dfbb9-9db2-4044-89b6-4399d061335d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.407187] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271739, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.410919] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56239c15-dff3-4f27-9adc-125a1989ec1e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.427014] env[62235]: DEBUG nova.virt.block_device [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Updating existing volume attachment record: a565dc29-1d1b-43f7-90e5-4dfe4ce95269 {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 853.529970] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e7df6081-6b2b-4a5f-8cb6-0228c4477671 tempest-ServerRescueTestJSONUnderV235-1369972033 tempest-ServerRescueTestJSONUnderV235-1369972033-project-member] Lock "876b7602-ccf0-4aed-90e2-d07fbf548fb6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.178s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.659613] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.659613] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.659988] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.660897] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.661169] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.664470] env[62235]: INFO nova.compute.manager [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Terminating instance [ 853.669572] env[62235]: DEBUG nova.compute.manager [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 853.669817] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.670944] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37229b4c-dea9-477f-b4e7-689edc07a4c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.681858] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.681858] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a8d1327-cb82-4829-a01d-593c2bae0b1e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.690297] env[62235]: DEBUG oslo_vmware.api [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 853.690297] env[62235]: value = "task-1271742" [ 853.690297] env[62235]: _type = "Task" [ 853.690297] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.704401] env[62235]: DEBUG oslo_vmware.api [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.800131] env[62235]: DEBUG nova.network.neutron [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.884926] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271738, 'name': ReconfigVM_Task, 'duration_secs': 0.686216} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.885377] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 15d41c0f-045d-4e1c-88b5-1841f99f2e74/15d41c0f-045d-4e1c-88b5-1841f99f2e74.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.889204] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddc9c048-f469-46bb-9308-71ee1e0a5371 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.895313] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 853.895313] env[62235]: value = "task-1271744" [ 853.895313] env[62235]: _type = "Task" [ 853.895313] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.907523] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271739, 'name': ReconfigVM_Task, 'duration_secs': 0.554723} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.908538] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Reconfigured VM instance instance-00000048 to attach disk [datastore1] b7801744-5552-49b8-a95d-41b67bdf2380/b7801744-5552-49b8-a95d-41b67bdf2380.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.909492] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2636fb34-ac63-49b0-bed0-0ee975303762 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.915112] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271744, 'name': Rename_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.921185] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 853.921185] env[62235]: value = "task-1271745" [ 853.921185] env[62235]: _type = "Task" [ 853.921185] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.930747] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271745, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.997889] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca0ac648-41f9-4068-ae6c-18b76338a8bc tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.999130] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.623s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.999373] env[62235]: DEBUG nova.objects.instance [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lazy-loading 'resources' on Instance uuid 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.112362] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "507fa559-1889-4f6e-81ee-93eb14d134b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.112778] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.178786] env[62235]: DEBUG nova.network.neutron [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Updating instance_info_cache with network_info: [{"id": "4c773abd-ae23-4075-aa4d-efb7b61a598c", "address": "fa:16:3e:a1:0e:8b", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c773abd-ae", "ovs_interfaceid": "4c773abd-ae23-4075-aa4d-efb7b61a598c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.200864] env[62235]: DEBUG oslo_vmware.api [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271742, 'name': PowerOffVM_Task, 'duration_secs': 0.190466} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.201242] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.204583] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.204583] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-848e2186-a712-4a01-8810-1c3d6d13f761 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.223782] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.224095] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.224267] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.224394] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.224563] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.227237] env[62235]: INFO nova.compute.manager [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Terminating instance [ 854.230423] env[62235]: DEBUG nova.compute.manager [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.230423] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 854.232685] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253f03f9-41ae-43bd-9a7b-1edd94e5d3a3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.239162] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 854.239476] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c90308b-8b3c-41e8-9019-a9798151e31b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.247257] env[62235]: DEBUG oslo_vmware.api [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 854.247257] env[62235]: value = "task-1271747" [ 854.247257] env[62235]: _type = "Task" [ 854.247257] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.257612] env[62235]: DEBUG oslo_vmware.api [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271747, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.340948] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.341075] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.341379] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Deleting the datastore file [datastore1] b126a2c9-62e9-49a1-91ae-7a703e5fa490 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.341464] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5317759e-9cbb-4ccf-8078-ca23cc692300 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.348405] env[62235]: DEBUG oslo_vmware.api [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for the task: (returnval){ [ 854.348405] env[62235]: value = "task-1271748" [ 854.348405] env[62235]: _type = "Task" [ 854.348405] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.356784] env[62235]: DEBUG oslo_vmware.api [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.408468] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271744, 'name': Rename_Task, 'duration_secs': 0.16222} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.408991] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.409427] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e2577b0-edd2-4f19-a631-547bff15c840 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.420424] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 854.420424] env[62235]: value = "task-1271749" [ 854.420424] env[62235]: _type = "Task" [ 854.420424] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.432258] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271745, 'name': Rename_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.435761] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271749, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.615592] env[62235]: DEBUG nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 854.681519] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "refresh_cache-c571a362-9b55-42d3-a019-1695c1caf248" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.682176] env[62235]: DEBUG nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Instance network_info: |[{"id": "4c773abd-ae23-4075-aa4d-efb7b61a598c", "address": "fa:16:3e:a1:0e:8b", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c773abd-ae", "ovs_interfaceid": "4c773abd-ae23-4075-aa4d-efb7b61a598c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.682594] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:0e:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c773abd-ae23-4075-aa4d-efb7b61a598c', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.690814] env[62235]: DEBUG oslo.service.loopingcall [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.691474] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.692335] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b67362ae-aa7a-4e8a-a386-ae50413bbd35 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.718459] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.718459] env[62235]: value = "task-1271750" [ 854.718459] env[62235]: _type = "Task" [ 854.718459] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.726650] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271750, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.757974] env[62235]: DEBUG oslo_vmware.api [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271747, 'name': PowerOffVM_Task, 'duration_secs': 0.372135} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.759084] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.759084] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.759084] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9f33740-05cc-4b6c-a0d9-d2448a3fafc5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.842495] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c564c3-3881-4eae-b93d-f74a9e74e33c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.853779] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644e6fbd-fd0f-4e83-8b26-8ba01d522f1a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.862310] env[62235]: DEBUG oslo_vmware.api [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Task: {'id': task-1271748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191719} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.889539] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.889828] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 854.890034] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 854.890303] env[62235]: INFO nova.compute.manager [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Took 1.22 seconds to destroy the instance on the hypervisor. [ 854.890642] env[62235]: DEBUG oslo.service.loopingcall [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.893664] env[62235]: DEBUG nova.compute.manager [-] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.893797] env[62235]: DEBUG nova.network.neutron [-] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.896579] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efc299e-ef03-409a-927b-6377bdec3306 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.899490] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.899641] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.899743] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleting the datastore file [datastore1] 00825c9e-b398-4edc-a5ed-e4b5b476af1c {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.903211] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70f09a79-4926-47d8-8ddd-641a9b030a3e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.905424] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "44ed8739-ad20-41a3-abe1-38176c1bee25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.905654] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.917562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0688cde-ab7b-4fbf-bde0-47cd7a152855 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.923887] env[62235]: DEBUG oslo_vmware.api [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 854.923887] env[62235]: value = "task-1271752" [ 854.923887] env[62235]: _type = "Task" [ 854.923887] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.945086] env[62235]: DEBUG nova.compute.provider_tree [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.953082] env[62235]: DEBUG oslo_vmware.api [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271749, 'name': PowerOnVM_Task, 'duration_secs': 0.511316} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.962947] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 854.963199] env[62235]: INFO nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Took 7.99 seconds to spawn the instance on the hypervisor. [ 854.963399] env[62235]: DEBUG nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.964138] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271745, 'name': Rename_Task, 'duration_secs': 0.681715} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.965483] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f892168-c79b-4de0-ad63-9e2fb30c5414 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.968482] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.972388] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed20215a-266c-4235-af45-e3a1dbe85be7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.974154] env[62235]: DEBUG oslo_vmware.api [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271752, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.984470] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 854.984470] env[62235]: value = "task-1271753" [ 854.984470] env[62235]: _type = "Task" [ 854.984470] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.995176] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.035889] env[62235]: DEBUG nova.compute.manager [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Received event network-vif-plugged-4c773abd-ae23-4075-aa4d-efb7b61a598c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.036168] env[62235]: DEBUG oslo_concurrency.lockutils [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] Acquiring lock "c571a362-9b55-42d3-a019-1695c1caf248-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.036425] env[62235]: DEBUG oslo_concurrency.lockutils [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] Lock "c571a362-9b55-42d3-a019-1695c1caf248-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.036683] env[62235]: DEBUG oslo_concurrency.lockutils [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] Lock "c571a362-9b55-42d3-a019-1695c1caf248-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.036914] env[62235]: DEBUG nova.compute.manager [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] No waiting events found dispatching network-vif-plugged-4c773abd-ae23-4075-aa4d-efb7b61a598c {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.037230] env[62235]: WARNING nova.compute.manager [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Received unexpected event network-vif-plugged-4c773abd-ae23-4075-aa4d-efb7b61a598c for instance with vm_state building and task_state spawning. [ 855.037547] env[62235]: DEBUG nova.compute.manager [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Received event network-changed-4c773abd-ae23-4075-aa4d-efb7b61a598c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.037547] env[62235]: DEBUG nova.compute.manager [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Refreshing instance network info cache due to event network-changed-4c773abd-ae23-4075-aa4d-efb7b61a598c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.037921] env[62235]: DEBUG oslo_concurrency.lockutils [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] Acquiring lock "refresh_cache-c571a362-9b55-42d3-a019-1695c1caf248" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.038215] env[62235]: DEBUG oslo_concurrency.lockutils [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] Acquired lock "refresh_cache-c571a362-9b55-42d3-a019-1695c1caf248" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.038476] env[62235]: DEBUG nova.network.neutron [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Refreshing network info cache for port 4c773abd-ae23-4075-aa4d-efb7b61a598c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.143302] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.228538] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271750, 'name': CreateVM_Task, 'duration_secs': 0.410752} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.228761] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.229517] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.230306] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.230306] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.230482] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b39cc4c4-3993-43d9-8a0d-71a89f027878 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.239276] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 855.239276] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f983f8-e356-aec3-1ee1-dee6474f844c" [ 855.239276] env[62235]: _type = "Task" [ 855.239276] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.249261] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f983f8-e356-aec3-1ee1-dee6474f844c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.411146] env[62235]: DEBUG nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.435511] env[62235]: DEBUG oslo_vmware.api [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271752, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167336} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.436031] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.436380] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 855.436710] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.437083] env[62235]: INFO nova.compute.manager [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Took 1.21 seconds to destroy the instance on the hypervisor. [ 855.437684] env[62235]: DEBUG oslo.service.loopingcall [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.438550] env[62235]: DEBUG nova.compute.manager [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.438550] env[62235]: DEBUG nova.network.neutron [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 855.450971] env[62235]: DEBUG nova.scheduler.client.report [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.499636] env[62235]: INFO nova.compute.manager [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Took 27.23 seconds to build instance. [ 855.500748] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.741252] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0b005bd8-498a-459a-8339-f08f32647fa4 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.741614] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0b005bd8-498a-459a-8339-f08f32647fa4 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.741845] env[62235]: DEBUG nova.objects.instance [None req-0b005bd8-498a-459a-8339-f08f32647fa4 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'flavor' on Instance uuid ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.752491] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.752743] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Processing image 4b23c73d-e39b-4f1a-8817-627621a5ee39 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.752981] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39/4b23c73d-e39b-4f1a-8817-627621a5ee39.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.753149] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39/4b23c73d-e39b-4f1a-8817-627621a5ee39.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.753340] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.753594] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a58991b3-559e-4e22-b858-cfda79d94071 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.765958] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.766187] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.766912] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aab40510-c7f1-40d9-b5a8-8d03e81cef76 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.773166] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 855.773166] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523d026b-440f-24f3-4fe2-8734909117af" [ 855.773166] env[62235]: _type = "Task" [ 855.773166] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.784731] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523d026b-440f-24f3-4fe2-8734909117af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.942089] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.959616] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.960s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.967345] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.824s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.969742] env[62235]: INFO nova.compute.claims [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.973680] env[62235]: DEBUG nova.network.neutron [-] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.998599] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.002795] env[62235]: INFO nova.scheduler.client.report [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Deleted allocations for instance 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9 [ 856.004216] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f11ae46d-3d31-4893-844c-1833b0589406 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.744s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.045906] env[62235]: DEBUG nova.network.neutron [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Updated VIF entry in instance network info cache for port 4c773abd-ae23-4075-aa4d-efb7b61a598c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 856.046298] env[62235]: DEBUG nova.network.neutron [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Updating instance_info_cache with network_info: [{"id": "4c773abd-ae23-4075-aa4d-efb7b61a598c", "address": "fa:16:3e:a1:0e:8b", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c773abd-ae", "ovs_interfaceid": "4c773abd-ae23-4075-aa4d-efb7b61a598c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.155534] env[62235]: DEBUG nova.network.neutron [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.186641] env[62235]: DEBUG nova.compute.manager [req-854f75b4-6e67-4832-9343-3e7631961601 req-668afa52-e5df-4038-87ac-321eceb609a3 service nova] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Received event network-vif-deleted-bfd2c6bf-a775-49d3-8e77-b851443ec66b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.196657] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.196912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.197118] env[62235]: INFO nova.compute.manager [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Shelving [ 856.248842] env[62235]: DEBUG nova.objects.instance [None req-0b005bd8-498a-459a-8339-f08f32647fa4 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'pci_requests' on Instance uuid ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.282718] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Preparing fetch location {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 856.282988] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Fetch image to [datastore1] OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222/OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222.vmdk {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 856.283198] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Downloading stream optimized image 4b23c73d-e39b-4f1a-8817-627621a5ee39 to [datastore1] OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222/OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222.vmdk on the data store datastore1 as vApp {{(pid=62235) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 856.283463] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Downloading image file data 4b23c73d-e39b-4f1a-8817-627621a5ee39 to the ESX as VM named 'OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222' {{(pid=62235) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 856.366570] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 856.366570] env[62235]: value = "resgroup-9" [ 856.366570] env[62235]: _type = "ResourcePool" [ 856.366570] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 856.366835] env[62235]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9174dad8-31ae-40e3-b63e-5a935054262d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.389066] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lease: (returnval){ [ 856.389066] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52326de3-923b-cbf6-28f1-824a2ff05cb0" [ 856.389066] env[62235]: _type = "HttpNfcLease" [ 856.389066] env[62235]: } obtained for vApp import into resource pool (val){ [ 856.389066] env[62235]: value = "resgroup-9" [ 856.389066] env[62235]: _type = "ResourcePool" [ 856.389066] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 856.389336] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the lease: (returnval){ [ 856.389336] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52326de3-923b-cbf6-28f1-824a2ff05cb0" [ 856.389336] env[62235]: _type = "HttpNfcLease" [ 856.389336] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 856.397712] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 856.397712] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52326de3-923b-cbf6-28f1-824a2ff05cb0" [ 856.397712] env[62235]: _type = "HttpNfcLease" [ 856.397712] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 856.480308] env[62235]: INFO nova.compute.manager [-] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Took 1.59 seconds to deallocate network for instance. [ 856.497328] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.515798] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d604c449-8290-44ad-968b-3f089a4d1538 tempest-SecurityGroupsTestJSON-1692744291 tempest-SecurityGroupsTestJSON-1692744291-project-member] Lock "9f5fcb83-bb6f-4815-b842-a37b2dbf26e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.604s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.549038] env[62235]: DEBUG oslo_concurrency.lockutils [req-05d97e42-157f-433e-850c-533c3e9bf918 req-916be0d3-bdc0-4792-8e57-0f1ab9ee02b4 service nova] Releasing lock "refresh_cache-c571a362-9b55-42d3-a019-1695c1caf248" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.660929] env[62235]: INFO nova.compute.manager [-] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Took 1.22 seconds to deallocate network for instance. [ 856.706906] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 856.707242] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed39cb45-0606-4821-b00c-e5876c640e6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.714524] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 856.714524] env[62235]: value = "task-1271756" [ 856.714524] env[62235]: _type = "Task" [ 856.714524] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.723711] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271756, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.753042] env[62235]: DEBUG nova.objects.base [None req-0b005bd8-498a-459a-8339-f08f32647fa4 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 856.753042] env[62235]: DEBUG nova.network.neutron [None req-0b005bd8-498a-459a-8339-f08f32647fa4 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 856.869059] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0b005bd8-498a-459a-8339-f08f32647fa4 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.127s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.898586] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 856.898586] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52326de3-923b-cbf6-28f1-824a2ff05cb0" [ 856.898586] env[62235]: _type = "HttpNfcLease" [ 856.898586] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 856.988013] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.998011] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271753, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.167908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.176165] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99ce0ca-b8a4-45e4-b6df-cb949c78ba85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.189747] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7bdc55-e1b5-4fe4-aac0-edee60925409 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.226241] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5696f920-e338-488f-a573-f114aaa8e946 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.234998] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271756, 'name': PowerOffVM_Task, 'duration_secs': 0.205215} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.237761] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.238570] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906838d3-ec43-4386-8e5f-b1f5158d4d6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.241978] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1656f3ea-69d2-4717-8f0b-ad5faff1043f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.269385] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4569b8-3eee-49db-893a-d4c94a84fac9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.272554] env[62235]: DEBUG nova.compute.provider_tree [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.398103] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 857.398103] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52326de3-923b-cbf6-28f1-824a2ff05cb0" [ 857.398103] env[62235]: _type = "HttpNfcLease" [ 857.398103] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 857.398445] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 857.398445] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52326de3-923b-cbf6-28f1-824a2ff05cb0" [ 857.398445] env[62235]: _type = "HttpNfcLease" [ 857.398445] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 857.401260] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c61da3-26bd-4589-8a78-cca06edeffd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.412513] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a22cbd-f6d9-0ee9-56b3-b66f5bb7abfe/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 857.412583] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a22cbd-f6d9-0ee9-56b3-b66f5bb7abfe/disk-0.vmdk. {{(pid=62235) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 857.480318] env[62235]: DEBUG nova.compute.manager [req-998f6efe-54eb-4ed9-b2f8-a01204df4dec req-9bdad49a-fb3d-43d2-915b-c848da99da26 service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received event network-vif-deleted-90607cc9-dfc7-4f25-b1dd-ad81a02691f9 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.480318] env[62235]: DEBUG nova.compute.manager [req-998f6efe-54eb-4ed9-b2f8-a01204df4dec req-9bdad49a-fb3d-43d2-915b-c848da99da26 service nova] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Received event network-vif-deleted-eede8be1-1cbc-4791-802e-52193488eb27 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.486023] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f7b50306-3998-4434-8245-d7606ab2ab6b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.501019] env[62235]: DEBUG oslo_vmware.api [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271753, 'name': PowerOnVM_Task, 'duration_secs': 2.370151} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.501019] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.501019] env[62235]: INFO nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Took 12.91 seconds to spawn the instance on the hypervisor. [ 857.501019] env[62235]: DEBUG nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.501277] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2fce37-8a32-4d7a-8d51-0b538a5de221 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.779776] env[62235]: DEBUG nova.scheduler.client.report [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.788359] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 857.788978] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2606fa57-fd58-40d3-9c00-344f1c6a829c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.799513] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 857.799513] env[62235]: value = "task-1271757" [ 857.799513] env[62235]: _type = "Task" [ 857.799513] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.814148] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271757, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.982727] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 857.982984] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273452', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'name': 'volume-083d43a4-743a-4f4c-872c-16b184c0c86c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9474de53-1964-4a52-8783-2bac5773b6c9', 'attached_at': '', 'detached_at': '', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'serial': '083d43a4-743a-4f4c-872c-16b184c0c86c'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 857.983916] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f315af9-d0c5-4e63-a0ac-3aa679043394 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.004884] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.009190] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.011188] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9db9af3-4dcf-4d4f-a4f5-71667f747617 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.028026] env[62235]: INFO nova.compute.manager [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Took 32.20 seconds to build instance. [ 858.058184] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] volume-083d43a4-743a-4f4c-872c-16b184c0c86c/volume-083d43a4-743a-4f4c-872c-16b184c0c86c.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.065615] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6109136f-36a2-4c18-9b12-de759f321742 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.089668] env[62235]: DEBUG oslo_vmware.api [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 858.089668] env[62235]: value = "task-1271758" [ 858.089668] env[62235]: _type = "Task" [ 858.089668] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.100557] env[62235]: DEBUG oslo_vmware.api [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271758, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.290744] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.291581] env[62235]: DEBUG nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.299114] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.356s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.300271] env[62235]: INFO nova.compute.claims [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.308010] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Completed reading data from the image iterator. {{(pid=62235) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 858.308143] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a22cbd-f6d9-0ee9-56b3-b66f5bb7abfe/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 858.309020] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f2e197-17b4-4812-b91e-547cfd773625 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.316195] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a22cbd-f6d9-0ee9-56b3-b66f5bb7abfe/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 858.317849] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a22cbd-f6d9-0ee9-56b3-b66f5bb7abfe/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 858.319672] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d201c317-a7dd-4c38-b3ce-9dec588a3b20 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.321716] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271757, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.521921] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.523496] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 858.523496] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 858.531566] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5afd3321-d5e4-40bf-90cc-a095a325f81c tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.714s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.544799] env[62235]: DEBUG oslo_vmware.rw_handles [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a22cbd-f6d9-0ee9-56b3-b66f5bb7abfe/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 858.545075] env[62235]: INFO nova.virt.vmwareapi.images [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Downloaded image file data 4b23c73d-e39b-4f1a-8817-627621a5ee39 [ 858.546118] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb72e24-9a75-45fe-a802-4f6af8f8a5ed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.565164] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee314a06-6d60-4d61-933a-a2d4cc1af193 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.600809] env[62235]: DEBUG oslo_vmware.api [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271758, 'name': ReconfigVM_Task, 'duration_secs': 0.507721} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.604016] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Reconfigured VM instance instance-00000047 to attach disk [datastore1] volume-083d43a4-743a-4f4c-872c-16b184c0c86c/volume-083d43a4-743a-4f4c-872c-16b184c0c86c.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.606585] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3679ce6-b36c-4174-b0dc-45d682b15184 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.621841] env[62235]: DEBUG oslo_vmware.api [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 858.621841] env[62235]: value = "task-1271760" [ 858.621841] env[62235]: _type = "Task" [ 858.621841] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.632205] env[62235]: DEBUG oslo_vmware.api [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271760, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.634250] env[62235]: INFO nova.virt.vmwareapi.images [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] The imported VM was unregistered [ 858.637013] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Caching image {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 858.637371] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating directory with path [datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.639236] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01a88e19-2e2c-4770-aa83-279d5276509c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.652021] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Created directory with path [datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.652021] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222/OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222.vmdk to [datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39/4b23c73d-e39b-4f1a-8817-627621a5ee39.vmdk. {{(pid=62235) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 858.652021] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0f1ef53a-f672-4abb-8885-cd585fbb1df1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.657988] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 858.657988] env[62235]: value = "task-1271761" [ 858.657988] env[62235]: _type = "Task" [ 858.657988] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.666725] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271761, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.800036] env[62235]: DEBUG nova.compute.utils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.800525] env[62235]: DEBUG nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.800838] env[62235]: DEBUG nova.network.neutron [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 858.820919] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271757, 'name': CreateSnapshot_Task, 'duration_secs': 0.708863} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.821283] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 858.822198] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73f44b8-f5d1-4d7c-9cb8-9719381c9b4a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.844772] env[62235]: DEBUG nova.policy [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '989eb180171f496bba4ad0781938ea1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4569f2ad03314a39829b7ad2aa011da1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.935626] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.936076] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.936300] env[62235]: DEBUG nova.objects.instance [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'flavor' on Instance uuid ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.034029] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 859.038430] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 859.038728] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 859.084610] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.084771] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.084926] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 859.085114] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid 3ea7db36-ebde-4523-b803-ad121a00269b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.137172] env[62235]: DEBUG oslo_vmware.api [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271760, 'name': ReconfigVM_Task, 'duration_secs': 0.166623} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.137172] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273452', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'name': 'volume-083d43a4-743a-4f4c-872c-16b184c0c86c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9474de53-1964-4a52-8783-2bac5773b6c9', 'attached_at': '', 'detached_at': '', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'serial': '083d43a4-743a-4f4c-872c-16b184c0c86c'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 859.151730] env[62235]: DEBUG nova.network.neutron [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Successfully created port: 6b26570e-71cc-4c85-b5d6-d16ced84d829 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.171969] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271761, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.306024] env[62235]: DEBUG nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.342694] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 859.348022] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f821b869-791a-4bdc-af8e-9a0041d334f5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.357037] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 859.357037] env[62235]: value = "task-1271762" [ 859.357037] env[62235]: _type = "Task" [ 859.357037] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.367424] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271762, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.395794] env[62235]: DEBUG nova.compute.manager [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.395883] env[62235]: DEBUG nova.compute.manager [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing instance network info cache due to event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.396604] env[62235]: DEBUG oslo_concurrency.lockutils [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] Acquiring lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.396604] env[62235]: DEBUG oslo_concurrency.lockutils [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] Acquired lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.396604] env[62235]: DEBUG nova.network.neutron [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.570145] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe80b3e-fdca-4f3b-a5a0-678967e60b7e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.579365] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2153a4-e0d3-41d7-a6d3-05f4e9ae8035 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.617448] env[62235]: DEBUG nova.objects.instance [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'pci_requests' on Instance uuid ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.619537] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217fcca6-487b-4bdc-8265-26fb06b116ed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.629957] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5b195f-ed45-4317-932d-a0d3c211b3fb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.648384] env[62235]: DEBUG nova.compute.provider_tree [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.670552] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271761, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.868630] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271762, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.124778] env[62235]: DEBUG nova.objects.base [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 860.125025] env[62235]: DEBUG nova.network.neutron [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 860.156310] env[62235]: DEBUG nova.scheduler.client.report [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.177851] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271761, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.196622] env[62235]: DEBUG nova.objects.instance [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'flavor' on Instance uuid 9474de53-1964-4a52-8783-2bac5773b6c9 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.255212] env[62235]: DEBUG nova.policy [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.317914] env[62235]: DEBUG nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 860.352390] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.353037] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.353037] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.353218] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.353416] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.353578] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.353811] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.353981] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.354189] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.354359] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.354612] env[62235]: DEBUG nova.virt.hardware [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.355846] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cb78d6-4983-45b6-8823-b1bfb6f73aac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.371768] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ebe991-a169-4d56-8d29-f780602dbb8e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.375446] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271762, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.430304] env[62235]: DEBUG nova.network.neutron [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updated VIF entry in instance network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.430735] env[62235]: DEBUG nova.network.neutron [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updating instance_info_cache with network_info: [{"id": "aab67c7c-76aa-41fa-acba-76bec34f355c", "address": "fa:16:3e:88:57:a9", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab67c7c-76", "ovs_interfaceid": "aab67c7c-76aa-41fa-acba-76bec34f355c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.675784] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.676765] env[62235]: DEBUG nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 860.684912] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271761, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.686224] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.698s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.686522] env[62235]: DEBUG nova.objects.instance [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lazy-loading 'resources' on Instance uuid b126a2c9-62e9-49a1-91ae-7a703e5fa490 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.703037] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1b6d644-2d50-4c46-bde7-0e279457d8a6 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.375s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.750772] env[62235]: DEBUG nova.network.neutron [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Successfully updated port: 6b26570e-71cc-4c85-b5d6-d16ced84d829 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.870092] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271762, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.935540] env[62235]: DEBUG oslo_concurrency.lockutils [req-06ecb143-cefc-4d99-adc6-18fa4887092d req-d1320e5e-26e0-49fd-aec1-9de1732da9aa service nova] Releasing lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.107434] env[62235]: DEBUG nova.network.neutron [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Successfully created port: c7098635-f26b-4913-9096-7241e46f0a61 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.174328] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271761, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.493589} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.174643] env[62235]: INFO nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222/OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222.vmdk to [datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39/4b23c73d-e39b-4f1a-8817-627621a5ee39.vmdk. [ 861.174836] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Cleaning up location [datastore1] OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 861.175048] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_92e04001-8c97-474d-86f8-8a2c08b18222 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.175374] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d3bf663-b9ca-4a1b-adf6-8e5bbf9f5a76 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.183766] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 861.183766] env[62235]: value = "task-1271763" [ 861.183766] env[62235]: _type = "Task" [ 861.183766] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.192034] env[62235]: DEBUG nova.compute.utils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.194393] env[62235]: DEBUG nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 861.194596] env[62235]: DEBUG nova.network.neutron [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 861.204901] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271763, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.237645] env[62235]: DEBUG nova.policy [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b64f6d6696c042079cced4c83d161f6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4469550d32b4482a9d5e9244cbd681e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.254807] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "refresh_cache-507fa559-1889-4f6e-81ee-93eb14d134b2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.254948] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquired lock "refresh_cache-507fa559-1889-4f6e-81ee-93eb14d134b2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.255365] env[62235]: DEBUG nova.network.neutron [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.372117] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271762, 'name': CloneVM_Task, 'duration_secs': 1.963133} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.375735] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Created linked-clone VM from snapshot [ 861.380904] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e0f7f9-a9fe-4e25-82a5-8bac59c394fb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.385718] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Uploading image 0b74899a-987d-4936-af42-5d58c7e24984 {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 861.393490] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.417907] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 861.417907] env[62235]: value = "vm-273456" [ 861.417907] env[62235]: _type = "VirtualMachine" [ 861.417907] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 861.419508] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f5642d35-c54b-4242-b7a1-b757d081ef13 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.426506] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease: (returnval){ [ 861.426506] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c7b0a8-da0a-8fd6-9b3f-24e6d74cd9e9" [ 861.426506] env[62235]: _type = "HttpNfcLease" [ 861.426506] env[62235]: } obtained for exporting VM: (result){ [ 861.426506] env[62235]: value = "vm-273456" [ 861.426506] env[62235]: _type = "VirtualMachine" [ 861.426506] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 861.426772] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the lease: (returnval){ [ 861.426772] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c7b0a8-da0a-8fd6-9b3f-24e6d74cd9e9" [ 861.426772] env[62235]: _type = "HttpNfcLease" [ 861.426772] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 861.432998] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 861.432998] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c7b0a8-da0a-8fd6-9b3f-24e6d74cd9e9" [ 861.432998] env[62235]: _type = "HttpNfcLease" [ 861.432998] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 861.446420] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3a46d0-b692-4448-b1dd-39401016b932 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.453693] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a40b0d2-b2f5-483f-a25f-5042e0cb7cdb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.492383] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dded4b-b6c0-4d1c-82ae-6665891be396 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.492383] env[62235]: DEBUG nova.network.neutron [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Successfully created port: 3c9265cd-f975-407d-8f23-ab587b57e089 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.499129] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175bccfd-3f7c-4c75-9cca-62e33f8ddfcd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.512835] env[62235]: DEBUG nova.compute.provider_tree [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.694950] env[62235]: DEBUG nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.698632] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271763, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.047568} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.698632] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.698632] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39/4b23c73d-e39b-4f1a-8817-627621a5ee39.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.698632] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39/4b23c73d-e39b-4f1a-8817-627621a5ee39.vmdk to [datastore1] c571a362-9b55-42d3-a019-1695c1caf248/c571a362-9b55-42d3-a019-1695c1caf248.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.698632] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15da5f06-510e-4106-a751-ad4927307ab8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.705159] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 861.705159] env[62235]: value = "task-1271765" [ 861.705159] env[62235]: _type = "Task" [ 861.705159] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.715479] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.803167] env[62235]: DEBUG nova.network.neutron [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.896631] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.896999] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 861.897074] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.897221] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.897409] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.897578] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.897803] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.897941] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.898111] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 861.898233] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.936430] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 861.936430] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c7b0a8-da0a-8fd6-9b3f-24e6d74cd9e9" [ 861.936430] env[62235]: _type = "HttpNfcLease" [ 861.936430] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 861.936766] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 861.936766] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c7b0a8-da0a-8fd6-9b3f-24e6d74cd9e9" [ 861.936766] env[62235]: _type = "HttpNfcLease" [ 861.936766] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 861.937535] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d47771-1313-4f0d-ba94-7320ba40515f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.944980] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2068d-0af9-d0ef-d91a-8790bbb12a73/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 861.945186] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2068d-0af9-d0ef-d91a-8790bbb12a73/disk-0.vmdk for reading. {{(pid=62235) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 862.018077] env[62235]: DEBUG nova.scheduler.client.report [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.081996] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8f519afc-4cf1-4560-a622-1df4b6c98fc8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.133717] env[62235]: DEBUG nova.network.neutron [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Updating instance_info_cache with network_info: [{"id": "6b26570e-71cc-4c85-b5d6-d16ced84d829", "address": "fa:16:3e:de:ab:5b", "network": {"id": "2899a929-b8c9-4aad-aeb8-fe5104886842", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1600754524-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4569f2ad03314a39829b7ad2aa011da1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c92f34c-1dd7-4dc5-b8e8-f6c55cc5b4b8", "external-id": "nsx-vlan-transportzone-850", "segmentation_id": 850, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b26570e-71", "ovs_interfaceid": "6b26570e-71cc-4c85-b5d6-d16ced84d829", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.216018] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271765, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.401192] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.466334] env[62235]: DEBUG nova.compute.manager [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Received event network-changed-9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.466796] env[62235]: DEBUG nova.compute.manager [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Refreshing instance network info cache due to event network-changed-9af619cc-4e60-4097-96e6-b957861ba812. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.467178] env[62235]: DEBUG oslo_concurrency.lockutils [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] Acquiring lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.467464] env[62235]: DEBUG oslo_concurrency.lockutils [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] Acquired lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.467730] env[62235]: DEBUG nova.network.neutron [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Refreshing network info cache for port 9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.526210] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.529319] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.361s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.530086] env[62235]: DEBUG nova.objects.instance [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lazy-loading 'resources' on Instance uuid 00825c9e-b398-4edc-a5ed-e4b5b476af1c {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.550858] env[62235]: INFO nova.scheduler.client.report [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Deleted allocations for instance b126a2c9-62e9-49a1-91ae-7a703e5fa490 [ 862.637644] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Releasing lock "refresh_cache-507fa559-1889-4f6e-81ee-93eb14d134b2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.637644] env[62235]: DEBUG nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Instance network_info: |[{"id": "6b26570e-71cc-4c85-b5d6-d16ced84d829", "address": "fa:16:3e:de:ab:5b", "network": {"id": "2899a929-b8c9-4aad-aeb8-fe5104886842", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1600754524-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4569f2ad03314a39829b7ad2aa011da1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c92f34c-1dd7-4dc5-b8e8-f6c55cc5b4b8", "external-id": "nsx-vlan-transportzone-850", "segmentation_id": 850, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b26570e-71", "ovs_interfaceid": "6b26570e-71cc-4c85-b5d6-d16ced84d829", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.637644] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:ab:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c92f34c-1dd7-4dc5-b8e8-f6c55cc5b4b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b26570e-71cc-4c85-b5d6-d16ced84d829', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.648129] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Creating folder: Project (4569f2ad03314a39829b7ad2aa011da1). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.649246] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a555a92b-0da9-4592-8e8c-74987502694d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.661897] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Created folder: Project (4569f2ad03314a39829b7ad2aa011da1) in parent group-v273362. [ 862.664886] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Creating folder: Instances. Parent ref: group-v273457. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.664886] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab445be9-7f50-4111-9cc2-cb57b5c033af {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.675155] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Created folder: Instances in parent group-v273457. [ 862.675501] env[62235]: DEBUG oslo.service.loopingcall [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.675986] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.677308] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3da17e8-5152-4717-9c11-f99a5b16bac4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.701780] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.701780] env[62235]: value = "task-1271768" [ 862.701780] env[62235]: _type = "Task" [ 862.701780] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.709320] env[62235]: DEBUG nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 862.722612] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271768, 'name': CreateVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.727504] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271765, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.737971] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.738494] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.738732] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.739134] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.739362] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.739622] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.740220] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.740604] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.740897] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.741171] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.741626] env[62235]: DEBUG nova.virt.hardware [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.742642] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1da68a5-7551-44da-a5a1-745cdb150b02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.751801] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e698eb-983b-4a73-9d7e-c39601df5100 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.039110] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "9474de53-1964-4a52-8783-2bac5773b6c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.039638] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.039884] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "9474de53-1964-4a52-8783-2bac5773b6c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.040276] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.040627] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.043372] env[62235]: INFO nova.compute.manager [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Terminating instance [ 863.048982] env[62235]: DEBUG nova.compute.manager [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.049471] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.050371] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d25b288b-c755-4b27-9b3e-dd66396e6476 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.060390] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 863.060390] env[62235]: value = "task-1271769" [ 863.060390] env[62235]: _type = "Task" [ 863.060390] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.061270] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b13c0389-598c-4155-ad53-4464311236e0 tempest-ServersTestMultiNic-1593234806 tempest-ServersTestMultiNic-1593234806-project-member] Lock "b126a2c9-62e9-49a1-91ae-7a703e5fa490" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.401s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.086459] env[62235]: DEBUG nova.network.neutron [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Successfully updated port: 3c9265cd-f975-407d-8f23-ab587b57e089 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.097604] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.183860] env[62235]: DEBUG nova.network.neutron [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Successfully updated port: c7098635-f26b-4913-9096-7241e46f0a61 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.226757] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271768, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.234151] env[62235]: DEBUG nova.compute.manager [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.234151] env[62235]: DEBUG nova.compute.manager [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing instance network info cache due to event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.234151] env[62235]: DEBUG oslo_concurrency.lockutils [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] Acquiring lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.234352] env[62235]: DEBUG oslo_concurrency.lockutils [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] Acquired lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.234402] env[62235]: DEBUG nova.network.neutron [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.243168] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271765, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.331891] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9d445f-d75f-4e2d-a35e-a0786ad40231 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.345510] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e279faa9-b0ed-471e-93bd-95522548e54c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.391736] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa24e916-dd42-4c89-9576-2dc71a8a4c8f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.400164] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d5991b-08e2-4c1f-9257-af74d0ef0dc0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.415883] env[62235]: DEBUG nova.compute.provider_tree [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.515463] env[62235]: DEBUG nova.network.neutron [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updated VIF entry in instance network info cache for port 9af619cc-4e60-4097-96e6-b957861ba812. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.516894] env[62235]: DEBUG nova.network.neutron [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updating instance_info_cache with network_info: [{"id": "9af619cc-4e60-4097-96e6-b957861ba812", "address": "fa:16:3e:be:96:6b", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9af619cc-4e", "ovs_interfaceid": "9af619cc-4e60-4097-96e6-b957861ba812", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.574691] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271769, 'name': PowerOffVM_Task, 'duration_secs': 0.262837} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.575089] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.575316] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 863.575520] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273452', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'name': 'volume-083d43a4-743a-4f4c-872c-16b184c0c86c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9474de53-1964-4a52-8783-2bac5773b6c9', 'attached_at': '', 'detached_at': '', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'serial': '083d43a4-743a-4f4c-872c-16b184c0c86c'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 863.576389] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79061d38-bba6-4e07-9c55-701247c66fe2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.602573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.602573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.602573] env[62235]: DEBUG nova.network.neutron [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.607240] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2440baf-36b1-4826-a13a-4f5de501d153 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.615416] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc90455-4e92-42dc-89e9-fd264c187f7e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.645164] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755b00f2-89e9-4575-980e-abf7752584b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.664233] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] The volume has not been displaced from its original location: [datastore1] volume-083d43a4-743a-4f4c-872c-16b184c0c86c/volume-083d43a4-743a-4f4c-872c-16b184c0c86c.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 863.669985] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 863.670568] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fb6187f-4d74-49ff-9b7d-eea8b72b7adf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.689136] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.689136] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.689136] env[62235]: DEBUG nova.network.neutron [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.691255] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 863.691255] env[62235]: value = "task-1271770" [ 863.691255] env[62235]: _type = "Task" [ 863.691255] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.702362] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271770, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.714672] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271768, 'name': CreateVM_Task, 'duration_secs': 0.518899} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.718505] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.719809] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.719809] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.720227] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.720992] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa8afca0-7ac2-4b2e-bfa3-13b9d622a091 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.726488] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271765, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.730397] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 863.730397] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52002a4c-663d-586c-2c77-10ce185d8d74" [ 863.730397] env[62235]: _type = "Task" [ 863.730397] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.741660] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52002a4c-663d-586c-2c77-10ce185d8d74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.919632] env[62235]: DEBUG nova.scheduler.client.report [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.019532] env[62235]: DEBUG oslo_concurrency.lockutils [req-6c0ad34e-da9c-427b-8917-0f9348ed391c req-53095be2-9bd3-4d1d-9dd2-2dd27510bce3 service nova] Releasing lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.108419] env[62235]: DEBUG nova.network.neutron [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updated VIF entry in instance network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 864.109255] env[62235]: DEBUG nova.network.neutron [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updating instance_info_cache with network_info: [{"id": "aab67c7c-76aa-41fa-acba-76bec34f355c", "address": "fa:16:3e:88:57:a9", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab67c7c-76", "ovs_interfaceid": "aab67c7c-76aa-41fa-acba-76bec34f355c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.161728] env[62235]: DEBUG nova.network.neutron [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.205154] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271770, 'name': ReconfigVM_Task, 'duration_secs': 0.366648} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.205720] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 864.211059] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37f4590e-2f88-41cc-a7e6-3e64b04feff4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.230437] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271765, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.232053] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 864.232053] env[62235]: value = "task-1271771" [ 864.232053] env[62235]: _type = "Task" [ 864.232053] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.243816] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52002a4c-663d-586c-2c77-10ce185d8d74, 'name': SearchDatastore_Task, 'duration_secs': 0.022316} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.247680] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.247992] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.248379] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.248617] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.248872] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.249208] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271771, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.249566] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b1b1197-734c-4903-bac9-339c9b318d60 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.258436] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.258694] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.259899] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3647239-fc10-48f7-914a-d1e6fddeeedf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.265515] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 864.265515] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5224a4e1-a21c-a57a-bab6-c635544d1919" [ 864.265515] env[62235]: _type = "Task" [ 864.265515] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.274345] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5224a4e1-a21c-a57a-bab6-c635544d1919, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.275394] env[62235]: WARNING nova.network.neutron [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] 695860c9-9102-4053-ad1c-75a1c4ac4b12 already exists in list: networks containing: ['695860c9-9102-4053-ad1c-75a1c4ac4b12']. ignoring it [ 864.426810] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.897s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.430559] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.029s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.430747] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.430833] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 864.432159] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b04025-e6e8-4873-9fdd-793d77da0dd4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.441583] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4ed280-bc76-4cfa-b987-486a29cda9f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.458895] env[62235]: INFO nova.scheduler.client.report [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted allocations for instance 00825c9e-b398-4edc-a5ed-e4b5b476af1c [ 864.465816] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c90746c-77ce-4c00-9af5-3ffede3f447e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.477963] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f91a51-4a20-4010-bd83-52caa8283b26 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.512949] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180979MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 864.513067] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.513460] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.577800] env[62235]: DEBUG nova.network.neutron [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance_info_cache with network_info: [{"id": "3c9265cd-f975-407d-8f23-ab587b57e089", "address": "fa:16:3e:b3:9f:3c", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c9265cd-f9", "ovs_interfaceid": "3c9265cd-f975-407d-8f23-ab587b57e089", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.614606] env[62235]: DEBUG oslo_concurrency.lockutils [req-2a8fdf40-2532-478b-b21b-f6fb71f6ab54 req-0c3e719c-e791-4241-baba-fa39e486a820 service nova] Releasing lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.733378] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271765, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.588585} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.733678] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4b23c73d-e39b-4f1a-8817-627621a5ee39/4b23c73d-e39b-4f1a-8817-627621a5ee39.vmdk to [datastore1] c571a362-9b55-42d3-a019-1695c1caf248/c571a362-9b55-42d3-a019-1695c1caf248.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.734574] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4f8a1a-06ef-4ae5-a79c-5d54d6f94751 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.760320] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] c571a362-9b55-42d3-a019-1695c1caf248/c571a362-9b55-42d3-a019-1695c1caf248.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.763860] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d87e0c09-8069-4662-b604-b39322fe06c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.779711] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271771, 'name': ReconfigVM_Task, 'duration_secs': 0.249115} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.782613] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273452', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'name': 'volume-083d43a4-743a-4f4c-872c-16b184c0c86c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9474de53-1964-4a52-8783-2bac5773b6c9', 'attached_at': '', 'detached_at': '', 'volume_id': '083d43a4-743a-4f4c-872c-16b184c0c86c', 'serial': '083d43a4-743a-4f4c-872c-16b184c0c86c'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 864.782910] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.786805] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa486a54-cd9f-4545-8695-d47cc12a4be1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.792330] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 864.792330] env[62235]: value = "task-1271772" [ 864.792330] env[62235]: _type = "Task" [ 864.792330] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.797494] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.797834] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5224a4e1-a21c-a57a-bab6-c635544d1919, 'name': SearchDatastore_Task, 'duration_secs': 0.015379} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.800741] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87a0ca79-a8ff-47f4-a88c-839a732d3565 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.803163] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3c0fc89-a27f-40f9-a4c7-c6c9df3ffeb5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.808662] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 864.808662] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521807cf-48ad-8753-8a60-555e05ab4fe4" [ 864.808662] env[62235]: _type = "Task" [ 864.808662] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.812192] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271772, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.821636] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521807cf-48ad-8753-8a60-555e05ab4fe4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.876973] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.877331] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.877458] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleting the datastore file [datastore1] 9474de53-1964-4a52-8783-2bac5773b6c9 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.877842] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22acca9c-b709-4cf8-9a0e-ea9a84f0054b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.885687] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 864.885687] env[62235]: value = "task-1271774" [ 864.885687] env[62235]: _type = "Task" [ 864.885687] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.898442] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.915692] env[62235]: DEBUG nova.network.neutron [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c7098635-f26b-4913-9096-7241e46f0a61", "address": "fa:16:3e:8c:72:73", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7098635-f2", "ovs_interfaceid": "c7098635-f26b-4913-9096-7241e46f0a61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.972484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-54480b0e-1140-49ed-a83a-834574b99ac5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "00825c9e-b398-4edc-a5ed-e4b5b476af1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.748s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.080824] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.081096] env[62235]: DEBUG nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Instance network_info: |[{"id": "3c9265cd-f975-407d-8f23-ab587b57e089", "address": "fa:16:3e:b3:9f:3c", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c9265cd-f9", "ovs_interfaceid": "3c9265cd-f975-407d-8f23-ab587b57e089", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.081545] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:9f:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c9265cd-f975-407d-8f23-ab587b57e089', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.089389] env[62235]: DEBUG oslo.service.loopingcall [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.089646] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.089906] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65e55263-b406-47cd-96f3-0f995caa2b2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.112647] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.112647] env[62235]: value = "task-1271775" [ 865.112647] env[62235]: _type = "Task" [ 865.112647] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.120836] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271775, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.307777] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271772, 'name': ReconfigVM_Task, 'duration_secs': 0.480727} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.308119] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Reconfigured VM instance instance-0000004a to attach disk [datastore1] c571a362-9b55-42d3-a019-1695c1caf248/c571a362-9b55-42d3-a019-1695c1caf248.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.308850] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70f8a630-a9ac-4270-9829-7923ed3e74a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.317269] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 865.317269] env[62235]: value = "task-1271776" [ 865.317269] env[62235]: _type = "Task" [ 865.317269] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.324181] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521807cf-48ad-8753-8a60-555e05ab4fe4, 'name': SearchDatastore_Task, 'duration_secs': 0.015689} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.324853] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.325122] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 507fa559-1889-4f6e-81ee-93eb14d134b2/507fa559-1889-4f6e-81ee-93eb14d134b2.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 865.325385] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b1f40e3-3800-4ee0-b529-3c2e1e4cb625 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.330441] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271776, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.334867] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 865.334867] env[62235]: value = "task-1271777" [ 865.334867] env[62235]: _type = "Task" [ 865.334867] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.344221] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271777, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.396927] env[62235]: DEBUG oslo_vmware.api [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184742} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.397252] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.398184] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.398184] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.398184] env[62235]: INFO nova.compute.manager [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Took 2.35 seconds to destroy the instance on the hypervisor. [ 865.398441] env[62235]: DEBUG oslo.service.loopingcall [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.398441] env[62235]: DEBUG nova.compute.manager [-] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.398734] env[62235]: DEBUG nova.network.neutron [-] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.418775] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.419635] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.419913] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.420968] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e67e66-992d-4717-acdf-6e732050f4bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.438841] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.439155] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.439351] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.439544] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.439697] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.439853] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.440089] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.440274] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.440448] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.440621] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.440799] env[62235]: DEBUG nova.virt.hardware [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.447097] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfiguring VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 865.449786] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df0f3865-516b-4151-bb0f-d280d8ea7c9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.468933] env[62235]: DEBUG oslo_vmware.api [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 865.468933] env[62235]: value = "task-1271778" [ 865.468933] env[62235]: _type = "Task" [ 865.468933] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.477600] env[62235]: DEBUG oslo_vmware.api [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271778, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.548582] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3ea7db36-ebde-4523-b803-ad121a00269b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.548827] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 63e16fd4-3a5a-4231-9013-c6a01f0eab80 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.549019] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0e9182f8-81e1-4ca2-97b6-0066faba0b22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.549260] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance d165f549-1759-41bf-86fd-b77793bd49c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.549362] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ced2aa67-ce75-4af5-9c55-773835d05733 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.549549] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 06ec6b9c-bef0-4f96-a185-4315961de7f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.549728] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 80a803d1-c034-42b6-a15d-9fc3893953ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.549854] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 9474de53-1964-4a52-8783-2bac5773b6c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.549970] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b7801744-5552-49b8-a95d-41b67bdf2380 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.550102] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 15d41c0f-045d-4e1c-88b5-1841f99f2e74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.550254] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance c571a362-9b55-42d3-a019-1695c1caf248 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.550395] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 507fa559-1889-4f6e-81ee-93eb14d134b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.550513] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 44ed8739-ad20-41a3-abe1-38176c1bee25 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 865.550726] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 865.550899] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 865.607445] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Received event network-vif-plugged-6b26570e-71cc-4c85-b5d6-d16ced84d829 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.607445] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Acquiring lock "507fa559-1889-4f6e-81ee-93eb14d134b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.607445] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.608239] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.608239] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] No waiting events found dispatching network-vif-plugged-6b26570e-71cc-4c85-b5d6-d16ced84d829 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.608442] env[62235]: WARNING nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Received unexpected event network-vif-plugged-6b26570e-71cc-4c85-b5d6-d16ced84d829 for instance with vm_state building and task_state spawning. [ 865.608565] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Received event network-changed-6b26570e-71cc-4c85-b5d6-d16ced84d829 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.608746] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Refreshing instance network info cache due to event network-changed-6b26570e-71cc-4c85-b5d6-d16ced84d829. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.608980] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Acquiring lock "refresh_cache-507fa559-1889-4f6e-81ee-93eb14d134b2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.609150] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Acquired lock "refresh_cache-507fa559-1889-4f6e-81ee-93eb14d134b2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.609326] env[62235]: DEBUG nova.network.neutron [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Refreshing network info cache for port 6b26570e-71cc-4c85-b5d6-d16ced84d829 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.629736] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271775, 'name': CreateVM_Task, 'duration_secs': 0.375471} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.629736] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.629736] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.629912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.630154] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.630903] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c50e5ee-c7ff-4406-9cd6-6adccff651b0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.640857] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 865.640857] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529e4d4b-a31c-7a0a-fca4-1c89ca45ebb6" [ 865.640857] env[62235]: _type = "Task" [ 865.640857] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.653032] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529e4d4b-a31c-7a0a-fca4-1c89ca45ebb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.770181] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae03b67-823a-4b26-9332-bb90dd4f40a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.779352] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f653a5a1-05c8-4c66-b87e-be8ca6af2966 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.811930] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e9c1ce-9f6c-4ae5-93a4-3087d9b3814a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.823776] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558453aa-583c-4e59-a84b-362e7a1a42bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.833238] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271776, 'name': Rename_Task, 'duration_secs': 0.155043} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.841840] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.842401] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.843665] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3aec3a14-76b1-4482-a535-1f4304a2027d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.853845] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271777, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.855880] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 865.855880] env[62235]: value = "task-1271779" [ 865.855880] env[62235]: _type = "Task" [ 865.855880] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.866475] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271779, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.979155] env[62235]: DEBUG oslo_vmware.api [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271778, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.151177] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529e4d4b-a31c-7a0a-fca4-1c89ca45ebb6, 'name': SearchDatastore_Task, 'duration_secs': 0.019207} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.151543] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.151792] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.152047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.152213] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.152402] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.152685] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e65ac8b1-43e4-4099-abac-f4925b522e2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.161061] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.161384] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.162140] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e777ea8-9036-47dc-85b8-74fcf9a0e07f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.168514] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 866.168514] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526f3183-27d5-c85a-37cb-824d1d714392" [ 866.168514] env[62235]: _type = "Task" [ 866.168514] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.177655] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526f3183-27d5-c85a-37cb-824d1d714392, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.308634] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Received event network-changed-bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.308634] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Refreshing instance network info cache due to event network-changed-bd95632a-e453-409f-ab45-023167eaf726. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.308945] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Acquiring lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.309172] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Acquired lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.309287] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Refreshing network info cache for port bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.352218] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.365159] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271777, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597542} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.365993] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 507fa559-1889-4f6e-81ee-93eb14d134b2/507fa559-1889-4f6e-81ee-93eb14d134b2.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.366324] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.366661] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6dfc839f-1b2f-46cb-a51b-5d6af69dda1f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.373090] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271779, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.378946] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 866.378946] env[62235]: value = "task-1271780" [ 866.378946] env[62235]: _type = "Task" [ 866.378946] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.390965] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271780, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.427320] env[62235]: DEBUG nova.network.neutron [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Updated VIF entry in instance network info cache for port 6b26570e-71cc-4c85-b5d6-d16ced84d829. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 866.427883] env[62235]: DEBUG nova.network.neutron [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Updating instance_info_cache with network_info: [{"id": "6b26570e-71cc-4c85-b5d6-d16ced84d829", "address": "fa:16:3e:de:ab:5b", "network": {"id": "2899a929-b8c9-4aad-aeb8-fe5104886842", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1600754524-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4569f2ad03314a39829b7ad2aa011da1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c92f34c-1dd7-4dc5-b8e8-f6c55cc5b4b8", "external-id": "nsx-vlan-transportzone-850", "segmentation_id": 850, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b26570e-71", "ovs_interfaceid": "6b26570e-71cc-4c85-b5d6-d16ced84d829", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.480547] env[62235]: DEBUG oslo_vmware.api [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271778, 'name': ReconfigVM_Task, 'duration_secs': 0.758444} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.481063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.481338] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfigured VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 866.525203] env[62235]: DEBUG nova.network.neutron [-] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.682857] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526f3183-27d5-c85a-37cb-824d1d714392, 'name': SearchDatastore_Task, 'duration_secs': 0.013171} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.683897] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86c1d7c0-1f1b-445a-b5a9-75da8dc455bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.689641] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 866.689641] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dea3fe-1d54-f88f-bf75-dde57b69bed2" [ 866.689641] env[62235]: _type = "Task" [ 866.689641] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.698343] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dea3fe-1d54-f88f-bf75-dde57b69bed2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.756224] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "b7801744-5552-49b8-a95d-41b67bdf2380" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.760029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.760029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.760029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.760029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.764247] env[62235]: INFO nova.compute.manager [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Terminating instance [ 866.770434] env[62235]: DEBUG nova.compute.manager [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.770770] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.771997] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38d01a8-100a-437d-bee1-d188e9d9ba97 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.780311] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.780606] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a72ae6d3-6e28-49d8-8dd8-f03e9a7732b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.787308] env[62235]: DEBUG oslo_vmware.api [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 866.787308] env[62235]: value = "task-1271781" [ 866.787308] env[62235]: _type = "Task" [ 866.787308] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.796024] env[62235]: DEBUG oslo_vmware.api [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.858066] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 866.858354] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.345s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.869725] env[62235]: DEBUG oslo_vmware.api [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271779, 'name': PowerOnVM_Task, 'duration_secs': 0.847523} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.870399] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.870399] env[62235]: INFO nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Took 15.67 seconds to spawn the instance on the hypervisor. [ 866.870399] env[62235]: DEBUG nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.871254] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccce011-c763-4d5b-b489-5e8c74418aba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.891151] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271780, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079196} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.891151] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.891151] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419b9f91-a7e9-4502-9096-bc3d238b8a67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.919893] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 507fa559-1889-4f6e-81ee-93eb14d134b2/507fa559-1889-4f6e-81ee-93eb14d134b2.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.920550] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eeb2d368-3ece-48be-9551-435a80db4956 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.942505] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Releasing lock "refresh_cache-507fa559-1889-4f6e-81ee-93eb14d134b2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.942654] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Received event network-vif-plugged-3c9265cd-f975-407d-8f23-ab587b57e089 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.942768] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Acquiring lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.942976] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.943166] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.943342] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] No waiting events found dispatching network-vif-plugged-3c9265cd-f975-407d-8f23-ab587b57e089 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.943517] env[62235]: WARNING nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Received unexpected event network-vif-plugged-3c9265cd-f975-407d-8f23-ab587b57e089 for instance with vm_state building and task_state spawning. [ 866.943686] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Received event network-changed-3c9265cd-f975-407d-8f23-ab587b57e089 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.943848] env[62235]: DEBUG nova.compute.manager [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Refreshing instance network info cache due to event network-changed-3c9265cd-f975-407d-8f23-ab587b57e089. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.944046] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Acquiring lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.944390] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Acquired lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.944390] env[62235]: DEBUG nova.network.neutron [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Refreshing network info cache for port 3c9265cd-f975-407d-8f23-ab587b57e089 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.951581] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 866.951581] env[62235]: value = "task-1271782" [ 866.951581] env[62235]: _type = "Task" [ 866.951581] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.961388] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271782, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.986752] env[62235]: DEBUG oslo_concurrency.lockutils [None req-bd2e3476-4f7e-4299-ae34-61947fafe259 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.051s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.032088] env[62235]: INFO nova.compute.manager [-] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Took 1.63 seconds to deallocate network for instance. [ 867.200700] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dea3fe-1d54-f88f-bf75-dde57b69bed2, 'name': SearchDatastore_Task, 'duration_secs': 0.011797} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.201689] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updated VIF entry in instance network info cache for port bd95632a-e453-409f-ab45-023167eaf726. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 867.202236] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updating instance_info_cache with network_info: [{"id": "bd95632a-e453-409f-ab45-023167eaf726", "address": "fa:16:3e:ef:ab:a1", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd95632a-e4", "ovs_interfaceid": "bd95632a-e453-409f-ab45-023167eaf726", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.206590] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.206590] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 44ed8739-ad20-41a3-abe1-38176c1bee25/44ed8739-ad20-41a3-abe1-38176c1bee25.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.206590] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64ddf04c-da18-422a-9fd3-26fb57ebec3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.215370] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 867.215370] env[62235]: value = "task-1271783" [ 867.215370] env[62235]: _type = "Task" [ 867.215370] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.225467] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271783, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.299519] env[62235]: DEBUG oslo_vmware.api [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271781, 'name': PowerOffVM_Task, 'duration_secs': 0.347797} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.299871] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.300090] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.300384] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-264b274a-f5dd-46f8-bc39-ecd16576b1a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.369284] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.369626] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.370721] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Deleting the datastore file [datastore1] b7801744-5552-49b8-a95d-41b67bdf2380 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.370721] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b414a9d8-63a0-4a41-b206-f79f27ae89c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.377950] env[62235]: DEBUG oslo_vmware.api [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 867.377950] env[62235]: value = "task-1271785" [ 867.377950] env[62235]: _type = "Task" [ 867.377950] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.394875] env[62235]: DEBUG oslo_vmware.api [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.398982] env[62235]: INFO nova.compute.manager [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Took 38.45 seconds to build instance. [ 867.462959] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271782, 'name': ReconfigVM_Task, 'duration_secs': 0.467952} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.463288] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 507fa559-1889-4f6e-81ee-93eb14d134b2/507fa559-1889-4f6e-81ee-93eb14d134b2.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.463981] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-992db724-486b-48e5-aae7-4fe284f1c44b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.471191] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 867.471191] env[62235]: value = "task-1271786" [ 867.471191] env[62235]: _type = "Task" [ 867.471191] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.481397] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271786, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.598238] env[62235]: INFO nova.compute.manager [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Took 0.57 seconds to detach 1 volumes for instance. [ 867.706030] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Releasing lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.706229] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-vif-plugged-c7098635-f26b-4913-9096-7241e46f0a61 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.706490] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.706656] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.707207] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.707207] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] No waiting events found dispatching network-vif-plugged-c7098635-f26b-4913-9096-7241e46f0a61 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.707207] env[62235]: WARNING nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received unexpected event network-vif-plugged-c7098635-f26b-4913-9096-7241e46f0a61 for instance with vm_state active and task_state None. [ 867.707428] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-changed-c7098635-f26b-4913-9096-7241e46f0a61 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.707473] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing instance network info cache due to event network-changed-c7098635-f26b-4913-9096-7241e46f0a61. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.707702] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.707916] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.708171] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing network info cache for port c7098635-f26b-4913-9096-7241e46f0a61 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.728048] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271783, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.807768] env[62235]: DEBUG nova.network.neutron [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updated VIF entry in instance network info cache for port 3c9265cd-f975-407d-8f23-ab587b57e089. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 867.808324] env[62235]: DEBUG nova.network.neutron [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance_info_cache with network_info: [{"id": "3c9265cd-f975-407d-8f23-ab587b57e089", "address": "fa:16:3e:b3:9f:3c", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c9265cd-f9", "ovs_interfaceid": "3c9265cd-f975-407d-8f23-ab587b57e089", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.888054] env[62235]: DEBUG oslo_vmware.api [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.473846} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.888350] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.888546] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.888938] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.888938] env[62235]: INFO nova.compute.manager [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Took 1.12 seconds to destroy the instance on the hypervisor. [ 867.889177] env[62235]: DEBUG oslo.service.loopingcall [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.889365] env[62235]: DEBUG nova.compute.manager [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.889478] env[62235]: DEBUG nova.network.neutron [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.901261] env[62235]: DEBUG oslo_concurrency.lockutils [None req-01e15768-fa6c-4b2c-8308-ed5e62ca030f tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "c571a362-9b55-42d3-a019-1695c1caf248" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.971s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.983144] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271786, 'name': Rename_Task, 'duration_secs': 0.208457} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.984487] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.984487] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cc9fb93-fa39-4d2a-8fcc-fdb9152b9017 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.991662] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 867.991662] env[62235]: value = "task-1271787" [ 867.991662] env[62235]: _type = "Task" [ 867.991662] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.000827] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271787, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.108786] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.109036] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.109273] env[62235]: DEBUG nova.objects.instance [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'resources' on Instance uuid 9474de53-1964-4a52-8783-2bac5773b6c9 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.229867] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271783, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534934} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.231345] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 44ed8739-ad20-41a3-abe1-38176c1bee25/44ed8739-ad20-41a3-abe1-38176c1bee25.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.231801] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.234324] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e5f1fdd-892c-4f4d-a655-23f8d2209d66 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.243528] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 868.243528] env[62235]: value = "task-1271788" [ 868.243528] env[62235]: _type = "Task" [ 868.243528] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.253793] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271788, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.313947] env[62235]: DEBUG oslo_concurrency.lockutils [req-5f977055-31fa-4237-a4a0-da1cafcf2566 req-5be22b60-c920-4364-a85b-ec6a09c10549 service nova] Releasing lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.505304] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271787, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.598339] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updated VIF entry in instance network info cache for port c7098635-f26b-4913-9096-7241e46f0a61. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.598812] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c7098635-f26b-4913-9096-7241e46f0a61", "address": "fa:16:3e:8c:72:73", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7098635-f2", "ovs_interfaceid": "c7098635-f26b-4913-9096-7241e46f0a61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.612688] env[62235]: DEBUG nova.compute.manager [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Received event network-vif-deleted-7bae7814-0344-4573-93e3-27f411ada206 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.612903] env[62235]: DEBUG nova.compute.manager [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.613276] env[62235]: DEBUG nova.compute.manager [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing instance network info cache due to event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.613387] env[62235]: DEBUG oslo_concurrency.lockutils [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] Acquiring lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.613557] env[62235]: DEBUG oslo_concurrency.lockutils [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] Acquired lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.613730] env[62235]: DEBUG nova.network.neutron [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.739555] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "c571a362-9b55-42d3-a019-1695c1caf248" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.739831] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "c571a362-9b55-42d3-a019-1695c1caf248" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.740069] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "c571a362-9b55-42d3-a019-1695c1caf248-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.740280] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "c571a362-9b55-42d3-a019-1695c1caf248-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.740530] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "c571a362-9b55-42d3-a019-1695c1caf248-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.746037] env[62235]: INFO nova.compute.manager [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Terminating instance [ 868.750742] env[62235]: DEBUG nova.compute.manager [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.750952] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.752062] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82235593-3af7-4d67-99cc-453aa6561b85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.760937] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271788, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07726} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.765919] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.766211] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.768130] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b2d32a-d707-479e-a960-7a45d031c059 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.774458] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f03a5fd-fda1-4491-b833-498ab34d9edc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.798931] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 44ed8739-ad20-41a3-abe1-38176c1bee25/44ed8739-ad20-41a3-abe1-38176c1bee25.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.803186] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61381f68-f5c0-4f5e-adf7-007e7bd8df2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.818408] env[62235]: DEBUG oslo_vmware.api [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 868.818408] env[62235]: value = "task-1271789" [ 868.818408] env[62235]: _type = "Task" [ 868.818408] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.823865] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 868.823865] env[62235]: value = "task-1271790" [ 868.823865] env[62235]: _type = "Task" [ 868.823865] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.831673] env[62235]: DEBUG oslo_vmware.api [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271789, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.843743] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271790, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.881150] env[62235]: DEBUG nova.network.neutron [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.916479] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fb7b45-56f9-4ee7-a4f3-48b98351a52f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.925777] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fc51c2-3e8c-44b9-8394-0efd245bb21c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.963051] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17328be-3259-4855-87f1-7292443d364a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.971420] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7dafaa0-c939-4954-9ffb-f145c5d1ef92 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.986654] env[62235]: DEBUG nova.compute.provider_tree [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.005045] env[62235]: DEBUG oslo_vmware.api [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271787, 'name': PowerOnVM_Task, 'duration_secs': 0.697949} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.005857] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.006222] env[62235]: INFO nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Took 8.69 seconds to spawn the instance on the hypervisor. [ 869.006441] env[62235]: DEBUG nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.007379] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1519bc05-4982-4e6e-8eaa-f43f10394624 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.102016] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.102339] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Received event network-changed-bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.102529] env[62235]: DEBUG nova.compute.manager [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Refreshing instance network info cache due to event network-changed-bd95632a-e453-409f-ab45-023167eaf726. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.102848] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Acquiring lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.102941] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Acquired lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.103066] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Refreshing network info cache for port bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 869.316305] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.317253] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.342787] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271790, 'name': ReconfigVM_Task, 'duration_secs': 0.429343} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.343054] env[62235]: DEBUG oslo_vmware.api [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271789, 'name': PowerOffVM_Task, 'duration_secs': 0.278414} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.345384] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 44ed8739-ad20-41a3-abe1-38176c1bee25/44ed8739-ad20-41a3-abe1-38176c1bee25.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.346057] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.346244] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.347105] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-daea6683-30ff-4668-a58d-9c1e34e48466 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.348738] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-737732d8-2aee-4916-aa1c-11ab8ecf4c39 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.360421] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 869.360421] env[62235]: value = "task-1271791" [ 869.360421] env[62235]: _type = "Task" [ 869.360421] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.372891] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271791, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.384075] env[62235]: INFO nova.compute.manager [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Took 1.49 seconds to deallocate network for instance. [ 869.394133] env[62235]: DEBUG nova.network.neutron [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updated VIF entry in instance network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.394560] env[62235]: DEBUG nova.network.neutron [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updating instance_info_cache with network_info: [{"id": "aab67c7c-76aa-41fa-acba-76bec34f355c", "address": "fa:16:3e:88:57:a9", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab67c7c-76", "ovs_interfaceid": "aab67c7c-76aa-41fa-acba-76bec34f355c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.424937] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.425189] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.425387] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleting the datastore file [datastore1] c571a362-9b55-42d3-a019-1695c1caf248 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.426308] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b80a490-cdee-46a1-9748-12d364e10f54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.434744] env[62235]: DEBUG oslo_vmware.api [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 869.434744] env[62235]: value = "task-1271793" [ 869.434744] env[62235]: _type = "Task" [ 869.434744] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.443720] env[62235]: DEBUG oslo_vmware.api [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.490280] env[62235]: DEBUG nova.scheduler.client.report [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.526756] env[62235]: INFO nova.compute.manager [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Took 14.41 seconds to build instance. [ 869.633378] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.764117] env[62235]: DEBUG nova.network.neutron [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.824436] env[62235]: DEBUG nova.compute.utils [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.871688] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271791, 'name': Rename_Task, 'duration_secs': 0.188425} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.871688] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.873050] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d4f287e6-29cc-4429-a595-ebe57d0f9e79 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.881280] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 869.881280] env[62235]: value = "task-1271794" [ 869.881280] env[62235]: _type = "Task" [ 869.881280] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.891367] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.891651] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271794, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.900535] env[62235]: DEBUG oslo_concurrency.lockutils [req-25014d06-b90c-49b6-8682-eee5c0d89527 req-6ab958ad-3085-41cf-bc09-5545dd728017 service nova] Releasing lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.945647] env[62235]: DEBUG oslo_vmware.api [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182166} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.945960] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.946105] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.946433] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.946509] env[62235]: INFO nova.compute.manager [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Took 1.20 seconds to destroy the instance on the hypervisor. [ 869.946741] env[62235]: DEBUG oslo.service.loopingcall [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.946909] env[62235]: DEBUG nova.compute.manager [-] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.947038] env[62235]: DEBUG nova.network.neutron [-] [instance: c571a362-9b55-42d3-a019-1695c1caf248] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 869.995543] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.886s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.998532] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.106s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.998532] env[62235]: DEBUG nova.objects.instance [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lazy-loading 'resources' on Instance uuid b7801744-5552-49b8-a95d-41b67bdf2380 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.019094] env[62235]: INFO nova.scheduler.client.report [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted allocations for instance 9474de53-1964-4a52-8783-2bac5773b6c9 [ 870.028848] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3ec9965e-16aa-4860-86d1-71d79419aba3 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.916s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.131168] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.131475] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.131969] env[62235]: DEBUG nova.objects.instance [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'flavor' on Instance uuid ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.267380] env[62235]: DEBUG oslo_concurrency.lockutils [req-64603130-8d09-4254-968a-8068f334c3c1 req-4309b4d2-a7c6-49d4-bc8d-5d2a9ae3ee65 service nova] Releasing lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.327209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.393065] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271794, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.470550] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2068d-0af9-d0ef-d91a-8790bbb12a73/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 870.471547] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0222f4a9-d8e5-4c80-afd9-46b2e3648d31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.481830] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2068d-0af9-d0ef-d91a-8790bbb12a73/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 870.482016] env[62235]: ERROR oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2068d-0af9-d0ef-d91a-8790bbb12a73/disk-0.vmdk due to incomplete transfer. [ 870.482265] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e437ce7b-3223-4f1d-96df-1d54acefe8d9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.491655] env[62235]: DEBUG oslo_vmware.rw_handles [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2068d-0af9-d0ef-d91a-8790bbb12a73/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 870.491655] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Uploaded image 0b74899a-987d-4936-af42-5d58c7e24984 to the Glance image server {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 870.494125] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 870.494341] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7e0ebf76-dc3f-42aa-9e7d-e6083cb35378 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.505236] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 870.505236] env[62235]: value = "task-1271795" [ 870.505236] env[62235]: _type = "Task" [ 870.505236] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.519029] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271795, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.526310] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d03c5a70-659a-4e00-bee1-882bde64d07d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "9474de53-1964-4a52-8783-2bac5773b6c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.487s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.675639] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d097ea5-64b4-4d87-889b-4b3e3501bc9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.686145] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa90b65-0022-47bd-b095-a92e56bc0f77 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.720177] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f0120b-bd4e-418f-8659-9d6e89abfeca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.728894] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefaa569-797c-44f5-b0f3-c55bba7536a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.743414] env[62235]: DEBUG nova.compute.provider_tree [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.763279] env[62235]: DEBUG nova.objects.instance [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'pci_requests' on Instance uuid ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.896025] env[62235]: DEBUG oslo_vmware.api [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271794, 'name': PowerOnVM_Task, 'duration_secs': 0.548412} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.896025] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.896025] env[62235]: INFO nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Took 8.18 seconds to spawn the instance on the hypervisor. [ 870.896025] env[62235]: DEBUG nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.896025] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd8b736-38fa-4b49-becb-0f6f5843d084 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.018123] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271795, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.067186] env[62235]: DEBUG nova.network.neutron [-] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.185181] env[62235]: DEBUG nova.compute.manager [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.186058] env[62235]: DEBUG nova.compute.manager [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing instance network info cache due to event network-changed-aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.186058] env[62235]: DEBUG oslo_concurrency.lockutils [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] Acquiring lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.188754] env[62235]: DEBUG oslo_concurrency.lockutils [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] Acquired lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.188754] env[62235]: DEBUG nova.network.neutron [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Refreshing network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.250023] env[62235]: DEBUG nova.scheduler.client.report [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.265957] env[62235]: DEBUG nova.objects.base [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 871.266222] env[62235]: DEBUG nova.network.neutron [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 871.352470] env[62235]: DEBUG nova.policy [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.405749] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.405749] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.405749] env[62235]: INFO nova.compute.manager [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Attaching volume 3ee292e3-3b28-4582-8270-d1b850f04967 to /dev/sdb [ 871.424495] env[62235]: INFO nova.compute.manager [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Took 15.51 seconds to build instance. [ 871.469999] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cef2134-0254-49f5-b225-aee5a06a4664 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.481481] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51022ca1-74e9-4d0b-ab70-d0a0c204d0e4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.503852] env[62235]: DEBUG nova.virt.block_device [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updating existing volume attachment record: dd0712ab-95bb-4f8b-bc7d-62fc768b662c {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 871.519683] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271795, 'name': Destroy_Task, 'duration_secs': 0.934802} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.520063] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Destroyed the VM [ 871.520346] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 871.520617] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bcad785c-0987-4418-83e7-5f6f15d39738 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.529020] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 871.529020] env[62235]: value = "task-1271796" [ 871.529020] env[62235]: _type = "Task" [ 871.529020] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.539774] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271796, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.570127] env[62235]: INFO nova.compute.manager [-] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Took 1.62 seconds to deallocate network for instance. [ 871.752761] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.757254] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "b7801744-5552-49b8-a95d-41b67bdf2380" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.794340] env[62235]: INFO nova.scheduler.client.report [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Deleted allocations for instance b7801744-5552-49b8-a95d-41b67bdf2380 [ 871.860234] env[62235]: DEBUG nova.compute.manager [req-d6e713e7-57ae-4ab2-9469-851a6e2914f2 req-c8bf0d5e-e792-4616-aeeb-5d6e573839d0 service nova] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Received event network-vif-deleted-bd95632a-e453-409f-ab45-023167eaf726 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.860234] env[62235]: DEBUG nova.compute.manager [req-d6e713e7-57ae-4ab2-9469-851a6e2914f2 req-c8bf0d5e-e792-4616-aeeb-5d6e573839d0 service nova] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Received event network-vif-deleted-4c773abd-ae23-4075-aa4d-efb7b61a598c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.926712] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c922b03-3efc-4acf-84b7-b099e19cce87 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.021s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.037936] env[62235]: DEBUG nova.network.neutron [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updated VIF entry in instance network info cache for port aab67c7c-76aa-41fa-acba-76bec34f355c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.037936] env[62235]: DEBUG nova.network.neutron [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updating instance_info_cache with network_info: [{"id": "aab67c7c-76aa-41fa-acba-76bec34f355c", "address": "fa:16:3e:88:57:a9", "network": {"id": "ad4523d2-e39d-4b37-a3fc-471e4a417901", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1245633853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4456e079eae64f41b1596821d41bd275", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab67c7c-76", "ovs_interfaceid": "aab67c7c-76aa-41fa-acba-76bec34f355c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.044669] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271796, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.079298] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.079582] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.079879] env[62235]: DEBUG nova.objects.instance [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lazy-loading 'resources' on Instance uuid c571a362-9b55-42d3-a019-1695c1caf248 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.308153] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0e39fc4c-89fb-4744-bc6c-35511f2de13e tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.551s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.309333] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.552s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.312847] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.312847] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.312847] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.312847] env[62235]: INFO nova.compute.manager [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Terminating instance [ 872.314763] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.314763] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquired lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.314763] env[62235]: DEBUG nova.network.neutron [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.542099] env[62235]: DEBUG oslo_concurrency.lockutils [req-dc13a159-a739-4d3e-b01a-25db1400e056 req-b45e9c9a-8509-4fe9-a640-c4247337bdbc service nova] Releasing lock "refresh_cache-d165f549-1759-41bf-86fd-b77793bd49c0" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.549101] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271796, 'name': RemoveSnapshot_Task, 'duration_secs': 0.662947} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.549101] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 872.549552] env[62235]: DEBUG nova.compute.manager [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.550394] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6967d5-8d8c-4252-9bb9-84caa7b05c57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.797714] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a3b161-d6f4-4a16-871c-e7a15a2cf01d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.807536] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c71396-f9cd-4b04-b325-6bab58c0e465 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.840983] env[62235]: DEBUG nova.compute.utils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Can not refresh info_cache because instance was not found {{(pid=62235) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 872.843878] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3446915-2b51-40b7-9e4d-1b4db6ba0e2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.854546] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90487353-2716-40c6-bcd2-c32d4e308a8a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.870023] env[62235]: DEBUG nova.compute.provider_tree [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.872088] env[62235]: DEBUG nova.network.neutron [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.063480] env[62235]: INFO nova.compute.manager [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Shelve offloading [ 873.065278] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.065537] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d72ee1f5-1c00-4af8-98c6-4db6533f9dea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.074974] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 873.074974] env[62235]: value = "task-1271800" [ 873.074974] env[62235]: _type = "Task" [ 873.074974] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.085809] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 873.086088] env[62235]: DEBUG nova.compute.manager [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.087583] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d454cbc6-6183-4bc4-bfb8-907ae7aae97c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.094725] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.094856] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.095166] env[62235]: DEBUG nova.network.neutron [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.115696] env[62235]: DEBUG nova.network.neutron [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.376237] env[62235]: DEBUG nova.scheduler.client.report [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.397457] env[62235]: DEBUG nova.compute.manager [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.400053] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122453b9-29aa-4182-8d26-3962d96825c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.450800] env[62235]: DEBUG nova.network.neutron [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Successfully updated port: f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.619674] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Releasing lock "refresh_cache-b7801744-5552-49b8-a95d-41b67bdf2380" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.623339] env[62235]: DEBUG nova.compute.manager [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.623339] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.623339] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29d91924-def3-4b8a-959e-35b0330eaabf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.634670] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50e776b-7f2c-4d22-90df-943ad726ec85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.669022] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b7801744-5552-49b8-a95d-41b67bdf2380 could not be found. [ 873.669022] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.669022] env[62235]: INFO nova.compute.manager [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Took 0.05 seconds to destroy the instance on the hypervisor. [ 873.669022] env[62235]: DEBUG oslo.service.loopingcall [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.669022] env[62235]: DEBUG nova.compute.manager [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.669022] env[62235]: DEBUG nova.network.neutron [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.692813] env[62235]: DEBUG nova.network.neutron [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.883400] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.801s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.894152] env[62235]: DEBUG nova.network.neutron [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.910845] env[62235]: INFO nova.compute.manager [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] instance snapshotting [ 873.917228] env[62235]: INFO nova.scheduler.client.report [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted allocations for instance c571a362-9b55-42d3-a019-1695c1caf248 [ 873.917691] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692c4f51-e75e-4543-a202-27e538b4e62e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.951027] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f06566-ff83-4158-8a85-bbabb272e271 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.954242] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.954836] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.954836] env[62235]: DEBUG nova.network.neutron [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.195495] env[62235]: DEBUG nova.network.neutron [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.396200] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.420941] env[62235]: DEBUG nova.compute.manager [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-vif-plugged-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.420941] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.421266] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.421602] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.421602] env[62235]: DEBUG nova.compute.manager [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] No waiting events found dispatching network-vif-plugged-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.421782] env[62235]: WARNING nova.compute.manager [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received unexpected event network-vif-plugged-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 for instance with vm_state active and task_state None. [ 874.421944] env[62235]: DEBUG nova.compute.manager [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-changed-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.422045] env[62235]: DEBUG nova.compute.manager [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing instance network info cache due to event network-changed-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.422223] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.436791] env[62235]: DEBUG oslo_concurrency.lockutils [None req-61ec5009-73e4-4175-acc8-d8957e84f7a2 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "c571a362-9b55-42d3-a019-1695c1caf248" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.697s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.437945] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "dbc7c100-fa14-41f5-bed9-52702e8f5f69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.438201] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "dbc7c100-fa14-41f5-bed9-52702e8f5f69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.464597] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 874.464694] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e6f107cb-c06d-4143-89ac-180005a0976e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.474089] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 874.474089] env[62235]: value = "task-1271802" [ 874.474089] env[62235]: _type = "Task" [ 874.474089] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.483068] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271802, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.526883] env[62235]: WARNING nova.network.neutron [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] 695860c9-9102-4053-ad1c-75a1c4ac4b12 already exists in list: networks containing: ['695860c9-9102-4053-ad1c-75a1c4ac4b12']. ignoring it [ 874.527122] env[62235]: WARNING nova.network.neutron [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] 695860c9-9102-4053-ad1c-75a1c4ac4b12 already exists in list: networks containing: ['695860c9-9102-4053-ad1c-75a1c4ac4b12']. ignoring it [ 874.603476] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "d165f549-1759-41bf-86fd-b77793bd49c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.603861] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "d165f549-1759-41bf-86fd-b77793bd49c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.604167] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "d165f549-1759-41bf-86fd-b77793bd49c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.604382] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "d165f549-1759-41bf-86fd-b77793bd49c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.604562] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "d165f549-1759-41bf-86fd-b77793bd49c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.606690] env[62235]: INFO nova.compute.manager [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Terminating instance [ 874.608541] env[62235]: DEBUG nova.compute.manager [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.608742] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.609834] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d278313-4157-43bd-857a-9b7abb1f714a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.618362] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.618594] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b5adce6-de64-47d9-9b6c-addf2112f80f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.629949] env[62235]: DEBUG oslo_vmware.api [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 874.629949] env[62235]: value = "task-1271803" [ 874.629949] env[62235]: _type = "Task" [ 874.629949] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.640642] env[62235]: DEBUG oslo_vmware.api [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271803, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.699690] env[62235]: INFO nova.compute.manager [-] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Took 1.03 seconds to deallocate network for instance. [ 874.809023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "80a803d1-c034-42b6-a15d-9fc3893953ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.809023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.809023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "80a803d1-c034-42b6-a15d-9fc3893953ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.810922] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.003s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.811337] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.813657] env[62235]: INFO nova.compute.manager [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Terminating instance [ 874.815899] env[62235]: DEBUG nova.compute.manager [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.816223] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.817549] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06693b93-97c7-4217-a9d0-299a883c81ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.829447] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.829447] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d014d3f9-fa29-4edb-b8c1-0a2fde22cc43 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.840016] env[62235]: DEBUG oslo_vmware.api [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 874.840016] env[62235]: value = "task-1271804" [ 874.840016] env[62235]: _type = "Task" [ 874.840016] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.847136] env[62235]: DEBUG oslo_vmware.api [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271804, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.943022] env[62235]: DEBUG nova.compute.manager [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 874.986074] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271802, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.140325] env[62235]: DEBUG oslo_vmware.api [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271803, 'name': PowerOffVM_Task, 'duration_secs': 0.251841} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.140619] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.140792] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.141061] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce18a44a-dc9b-497e-9b27-bfaf1407c969 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.203953] env[62235]: INFO nova.compute.manager [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance disappeared during terminate [ 875.204282] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd6daa26-8609-4906-9f37-410ef7d4d470 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "b7801744-5552-49b8-a95d-41b67bdf2380" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.895s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.207423] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.208450] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eabec20-24a2-49ff-8b4d-71183de5dd69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.217419] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.218308] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17ef8ee5-d949-43d5-b2fb-2243834ab305 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.228837] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.229231] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.229537] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Deleting the datastore file [datastore2] d165f549-1759-41bf-86fd-b77793bd49c0 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.229906] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56f4b837-e09b-4e1f-9fe6-87ceab8524d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.238133] env[62235]: DEBUG oslo_vmware.api [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for the task: (returnval){ [ 875.238133] env[62235]: value = "task-1271807" [ 875.238133] env[62235]: _type = "Task" [ 875.238133] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.250217] env[62235]: DEBUG oslo_vmware.api [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271807, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.320024] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.320264] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.320610] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleting the datastore file [datastore2] 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.320726] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-034d6551-46d4-4879-a855-74f02841fecc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.327271] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 875.327271] env[62235]: value = "task-1271808" [ 875.327271] env[62235]: _type = "Task" [ 875.327271] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.335943] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271808, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.345637] env[62235]: DEBUG oslo_vmware.api [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271804, 'name': PowerOffVM_Task, 'duration_secs': 0.195706} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.345914] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.346116] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.346369] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18241a8c-1004-4c02-b895-7b050137ef38 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.389312] env[62235]: DEBUG nova.network.neutron [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c7098635-f26b-4913-9096-7241e46f0a61", "address": "fa:16:3e:8c:72:73", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7098635-f2", "ovs_interfaceid": "c7098635-f26b-4913-9096-7241e46f0a61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f80efd23-9c8c-4c33-8d50-e2bbbb4f8770", "address": "fa:16:3e:1c:4d:69", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80efd23-9c", "ovs_interfaceid": "f80efd23-9c8c-4c33-8d50-e2bbbb4f8770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.415983] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.416625] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.417038] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Deleting the datastore file [datastore1] 80a803d1-c034-42b6-a15d-9fc3893953ad {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.417437] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0591e8b-e1d6-4b89-b2d3-5af97d5bb040 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.426596] env[62235]: DEBUG oslo_vmware.api [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for the task: (returnval){ [ 875.426596] env[62235]: value = "task-1271810" [ 875.426596] env[62235]: _type = "Task" [ 875.426596] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.436122] env[62235]: DEBUG oslo_vmware.api [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.463799] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.464104] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.465728] env[62235]: INFO nova.compute.claims [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.484967] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271802, 'name': CreateSnapshot_Task, 'duration_secs': 0.669743} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.485295] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 875.486164] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0236c88a-af44-4df6-8008-8dce3a372a85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.749433] env[62235]: DEBUG oslo_vmware.api [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Task: {'id': task-1271807, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282321} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.749742] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.749911] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.750107] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.750377] env[62235]: INFO nova.compute.manager [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 875.750530] env[62235]: DEBUG oslo.service.loopingcall [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.750733] env[62235]: DEBUG nova.compute.manager [-] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.750830] env[62235]: DEBUG nova.network.neutron [-] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.839023] env[62235]: DEBUG oslo_vmware.api [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271808, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227668} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.839355] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.839593] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.839718] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.858306] env[62235]: INFO nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted allocations for instance 0e9182f8-81e1-4ca2-97b6-0066faba0b22 [ 875.892058] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.892842] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.893077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.894577] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.894772] env[62235]: DEBUG nova.network.neutron [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Refreshing network info cache for port f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.896509] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4578a8d-5281-4fcb-80f7-54b65d022375 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.917919] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.918224] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.918397] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.918604] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.918760] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.918926] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.919161] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.919333] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.919517] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.919695] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.919876] env[62235]: DEBUG nova.virt.hardware [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.930597] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfiguring VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 875.931446] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b3cfbfa-63d5-4ee7-ae2b-b654b458aaea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.955450] env[62235]: DEBUG oslo_vmware.api [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Task: {'id': task-1271810, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29209} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.955758] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.955953] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.956175] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.956382] env[62235]: INFO nova.compute.manager [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Took 1.14 seconds to destroy the instance on the hypervisor. [ 875.956617] env[62235]: DEBUG oslo.service.loopingcall [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.956859] env[62235]: DEBUG nova.compute.manager [-] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.956969] env[62235]: DEBUG nova.network.neutron [-] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.961071] env[62235]: DEBUG oslo_vmware.api [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 875.961071] env[62235]: value = "task-1271811" [ 875.961071] env[62235]: _type = "Task" [ 875.961071] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.970589] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.506s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.971529] env[62235]: DEBUG nova.compute.utils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Instance dbc7c100-fa14-41f5-bed9-52702e8f5f69 could not be found. {{(pid=62235) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 875.972859] env[62235]: DEBUG oslo_vmware.api [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271811, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.973616] env[62235]: DEBUG nova.compute.manager [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Instance disappeared during build. {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2513}} [ 875.973783] env[62235]: DEBUG nova.compute.manager [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Unplugging VIFs for instance {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 875.974047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-dbc7c100-fa14-41f5-bed9-52702e8f5f69" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.974208] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-dbc7c100-fa14-41f5-bed9-52702e8f5f69" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.974375] env[62235]: DEBUG nova.network.neutron [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 876.004879] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 876.005522] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e9249514-271f-4469-b408-55479a461ee4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.017602] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 876.017602] env[62235]: value = "task-1271812" [ 876.017602] env[62235]: _type = "Task" [ 876.017602] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.029548] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271812, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.365359] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.365359] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.365359] env[62235]: DEBUG nova.objects.instance [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'resources' on Instance uuid 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.471734] env[62235]: DEBUG oslo_vmware.api [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271811, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.477248] env[62235]: DEBUG nova.compute.utils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Can not refresh info_cache because instance was not found {{(pid=62235) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 876.492732] env[62235]: DEBUG nova.network.neutron [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.529741] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271812, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.566172] env[62235]: DEBUG nova.network.neutron [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.572044] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 876.572321] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273462', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'name': 'volume-3ee292e3-3b28-4582-8270-d1b850f04967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'serial': '3ee292e3-3b28-4582-8270-d1b850f04967'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 876.573476] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221b42b9-3514-4405-b702-c44582a090d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.596432] env[62235]: DEBUG nova.network.neutron [-] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.598418] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712e7979-f978-452d-ab37-6db1f8af2431 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.618902] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.619223] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.619442] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.619626] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.619798] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.628566] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] volume-3ee292e3-3b28-4582-8270-d1b850f04967/volume-3ee292e3-3b28-4582-8270-d1b850f04967.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.630772] env[62235]: INFO nova.compute.manager [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Terminating instance [ 876.633952] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48613896-0d99-4226-8d22-1fe24e2e0072 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.649483] env[62235]: DEBUG nova.compute.manager [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.649734] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.651184] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cf2f12-7924-4983-95bc-31b55c08b90e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.662096] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.663633] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6bc80a13-aefd-4f54-b5a6-f61bb139b058 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.666273] env[62235]: DEBUG oslo_vmware.api [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 876.666273] env[62235]: value = "task-1271813" [ 876.666273] env[62235]: _type = "Task" [ 876.666273] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.683069] env[62235]: DEBUG oslo_vmware.api [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271813, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.683069] env[62235]: DEBUG oslo_vmware.api [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 876.683069] env[62235]: value = "task-1271814" [ 876.683069] env[62235]: _type = "Task" [ 876.683069] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.697538] env[62235]: DEBUG oslo_vmware.api [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.867280] env[62235]: DEBUG nova.objects.instance [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'numa_topology' on Instance uuid 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.959594] env[62235]: DEBUG nova.network.neutron [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updated VIF entry in instance network info cache for port f80efd23-9c8c-4c33-8d50-e2bbbb4f8770. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.960164] env[62235]: DEBUG nova.network.neutron [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c7098635-f26b-4913-9096-7241e46f0a61", "address": "fa:16:3e:8c:72:73", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7098635-f2", "ovs_interfaceid": "c7098635-f26b-4913-9096-7241e46f0a61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f80efd23-9c8c-4c33-8d50-e2bbbb4f8770", "address": "fa:16:3e:1c:4d:69", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80efd23-9c", "ovs_interfaceid": "f80efd23-9c8c-4c33-8d50-e2bbbb4f8770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.972267] env[62235]: DEBUG oslo_vmware.api [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271811, 'name': ReconfigVM_Task, 'duration_secs': 0.880686} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.972795] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.973029] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfigured VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 877.030713] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271812, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.069722] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-dbc7c100-fa14-41f5-bed9-52702e8f5f69" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.070098] env[62235]: DEBUG nova.compute.manager [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62235) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 877.070376] env[62235]: DEBUG nova.compute.manager [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.070638] env[62235]: DEBUG nova.network.neutron [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 877.087166] env[62235]: DEBUG nova.network.neutron [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.101911] env[62235]: INFO nova.compute.manager [-] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Took 1.35 seconds to deallocate network for instance. [ 877.178279] env[62235]: DEBUG oslo_vmware.api [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271813, 'name': ReconfigVM_Task, 'duration_secs': 0.425227} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.178279] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfigured VM instance instance-00000042 to attach disk [datastore1] volume-3ee292e3-3b28-4582-8270-d1b850f04967/volume-3ee292e3-3b28-4582-8270-d1b850f04967.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.182839] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83b3c410-ef6a-4163-b093-c39e3ea7e1fb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.203606] env[62235]: DEBUG nova.compute.manager [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Stashing vm_state: active {{(pid=62235) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 877.211425] env[62235]: DEBUG oslo_vmware.api [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271814, 'name': PowerOffVM_Task, 'duration_secs': 0.203845} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.211734] env[62235]: DEBUG oslo_vmware.api [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 877.211734] env[62235]: value = "task-1271815" [ 877.211734] env[62235]: _type = "Task" [ 877.211734] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.212228] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.212407] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.212769] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c069845-a43a-4a2d-a3a0-7b0432ee0987 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.226167] env[62235]: DEBUG oslo_vmware.api [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271815, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.309256] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.309664] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.309864] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleting the datastore file [datastore2] 63e16fd4-3a5a-4231-9013-c6a01f0eab80 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.310193] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7588c9ab-47a4-4fe8-8441-8b5971c53acb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.318772] env[62235]: DEBUG oslo_vmware.api [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 877.318772] env[62235]: value = "task-1271817" [ 877.318772] env[62235]: _type = "Task" [ 877.318772] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.329037] env[62235]: DEBUG oslo_vmware.api [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.371370] env[62235]: DEBUG nova.objects.base [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Object Instance<0e9182f8-81e1-4ca2-97b6-0066faba0b22> lazy-loaded attributes: resources,numa_topology {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 877.466129] env[62235]: DEBUG oslo_concurrency.lockutils [req-5c9dfc88-7fed-462f-ac8c-01d3d2b56261 req-5f26e215-9834-40f3-8e68-5deb1205cca3 service nova] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.478215] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d74bb788-1b66-4f67-a925-5f271c50ceca tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.347s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.518734] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fe6bd9-9615-457d-9425-9cb2840b9bb3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.533248] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc56377-ac33-4e5f-85e7-9c13d13762be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.536856] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271812, 'name': CloneVM_Task, 'duration_secs': 1.326975} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.537164] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Created linked-clone VM from snapshot [ 877.538285] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b477bc31-adbe-43a5-80c3-7119456ba852 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.564678] env[62235]: DEBUG nova.network.neutron [-] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.571342] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d6c779-a515-4a83-8cd4-ed9e4b3556e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.574531] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Uploading image b1ebb36c-537b-4e31-9dfe-858f1f81662b {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 877.579794] env[62235]: DEBUG nova.compute.manager [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-vif-unplugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.580021] env[62235]: DEBUG oslo_concurrency.lockutils [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.580627] env[62235]: DEBUG oslo_concurrency.lockutils [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.580627] env[62235]: DEBUG oslo_concurrency.lockutils [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.580735] env[62235]: DEBUG nova.compute.manager [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] No waiting events found dispatching network-vif-unplugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.580809] env[62235]: WARNING nova.compute.manager [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received unexpected event network-vif-unplugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 for instance with vm_state shelved_offloaded and task_state None. [ 877.580971] env[62235]: DEBUG nova.compute.manager [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.581143] env[62235]: DEBUG nova.compute.manager [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing instance network info cache due to event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.581331] env[62235]: DEBUG oslo_concurrency.lockutils [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.581470] env[62235]: DEBUG oslo_concurrency.lockutils [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.581625] env[62235]: DEBUG nova.network.neutron [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.583800] env[62235]: INFO nova.compute.manager [-] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Took 1.63 seconds to deallocate network for instance. [ 877.591403] env[62235]: DEBUG nova.network.neutron [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.598230] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ad08ac-acc3-4df5-b462-fdd45f022d4c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.608864] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.619291] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 877.622360] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 877.622360] env[62235]: value = "vm-273464" [ 877.622360] env[62235]: _type = "VirtualMachine" [ 877.622360] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 877.623164] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8a0e017a-4609-49ac-b8e2-8936c97ae78d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.634172] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lease: (returnval){ [ 877.634172] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524f6fc6-34d3-aad3-a984-b753398948c8" [ 877.634172] env[62235]: _type = "HttpNfcLease" [ 877.634172] env[62235]: } obtained for exporting VM: (result){ [ 877.634172] env[62235]: value = "vm-273464" [ 877.634172] env[62235]: _type = "VirtualMachine" [ 877.634172] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 877.634172] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the lease: (returnval){ [ 877.634172] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524f6fc6-34d3-aad3-a984-b753398948c8" [ 877.634172] env[62235]: _type = "HttpNfcLease" [ 877.634172] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 877.642424] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.642424] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524f6fc6-34d3-aad3-a984-b753398948c8" [ 877.642424] env[62235]: _type = "HttpNfcLease" [ 877.642424] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 877.726341] env[62235]: DEBUG oslo_vmware.api [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271815, 'name': ReconfigVM_Task, 'duration_secs': 0.16016} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.727724] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.727724] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273462', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'name': 'volume-3ee292e3-3b28-4582-8270-d1b850f04967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'serial': '3ee292e3-3b28-4582-8270-d1b850f04967'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 877.777217] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.777731] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.829528] env[62235]: DEBUG oslo_vmware.api [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.456263} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.829806] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.830021] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 877.830262] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 877.830451] env[62235]: INFO nova.compute.manager [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Took 1.18 seconds to destroy the instance on the hypervisor. [ 877.830701] env[62235]: DEBUG oslo.service.loopingcall [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.830901] env[62235]: DEBUG nova.compute.manager [-] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.831019] env[62235]: DEBUG nova.network.neutron [-] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.092621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.103505] env[62235]: INFO nova.compute.manager [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: dbc7c100-fa14-41f5-bed9-52702e8f5f69] Took 1.03 seconds to deallocate network for instance. [ 878.111736] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "b5e4c6b2-4583-4a20-b570-99699063e172" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.112013] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "b5e4c6b2-4583-4a20-b570-99699063e172" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.145498] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 878.145498] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524f6fc6-34d3-aad3-a984-b753398948c8" [ 878.145498] env[62235]: _type = "HttpNfcLease" [ 878.145498] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 878.145874] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 878.145874] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524f6fc6-34d3-aad3-a984-b753398948c8" [ 878.145874] env[62235]: _type = "HttpNfcLease" [ 878.145874] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 878.146660] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e79ec9-bcf7-4053-8750-4fcafcf1f045 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.153182] env[62235]: ERROR nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [req-d9f595bc-af2f-42b5-aabc-7acbeb02c852] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d9f595bc-af2f-42b5-aabc-7acbeb02c852"}]} [ 878.161490] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c40dae-a5ed-cacb-db5d-0dae37bf58f7/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 878.161721] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c40dae-a5ed-cacb-db5d-0dae37bf58f7/disk-0.vmdk for reading. {{(pid=62235) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 878.228999] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 878.250452] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 878.250770] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 878.264248] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 878.268753] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7d7ad24b-067f-4028-aa40-844474ec5c28 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.281856] env[62235]: DEBUG nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 878.290218] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 878.453654] env[62235]: DEBUG nova.network.neutron [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updated VIF entry in instance network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.454084] env[62235]: DEBUG nova.network.neutron [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.497575] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d2a68c-4fc6-4e7c-9791-e8c50e8a0acb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.507765] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d08e4f1-d292-4a0c-9809-58ac86ea6d40 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.540460] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203990a2-cc77-433b-8ec2-87ce40449483 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.549481] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31daaf78-06e3-404f-898b-12189ece83e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.564658] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 878.618270] env[62235]: DEBUG nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 878.739688] env[62235]: DEBUG nova.network.neutron [-] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.780240] env[62235]: DEBUG nova.objects.instance [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lazy-loading 'flavor' on Instance uuid 06ec6b9c-bef0-4f96-a185-4315961de7f7 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.805908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.960749] env[62235]: DEBUG oslo_concurrency.lockutils [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.960749] env[62235]: DEBUG nova.compute.manager [req-19b1ae6a-c5b2-4988-8e93-57bdf7509497 req-93829306-8a4e-4086-bd74-b00a69e50080 service nova] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Received event network-vif-deleted-aab67c7c-76aa-41fa-acba-76bec34f355c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.084787] env[62235]: ERROR nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [req-df6eba7b-f451-4373-a17a-12f4bb1bd742] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-df6eba7b-f451-4373-a17a-12f4bb1bd742"}]} [ 879.103833] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 879.121271] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 879.121271] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 879.133756] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d697152f-dd2c-44bf-8bdc-559b5a25d320 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "dbc7c100-fa14-41f5-bed9-52702e8f5f69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 4.695s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.136241] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 879.149114] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.157940] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 879.230322] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-c7098635-f26b-4913-9096-7241e46f0a61" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.230588] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-c7098635-f26b-4913-9096-7241e46f0a61" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.244394] env[62235]: INFO nova.compute.manager [-] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Took 1.41 seconds to deallocate network for instance. [ 879.286421] env[62235]: DEBUG oslo_concurrency.lockutils [None req-daee8336-cf51-4a2a-a516-8dedacebb4d5 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.881s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.336977] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d308f8-5f62-4075-a410-98d840628d8c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.347091] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1f15d9-f702-4974-916e-ef132b6f461a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.386647] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723d8766-d317-4985-9708-6edf1d77a152 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.396675] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd556036-0513-4b65-9521-dd19b36f281a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.416611] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 879.593926] env[62235]: DEBUG nova.compute.manager [req-770072bd-2bc7-4217-952b-b66385aa53ef req-183652e9-eb42-43fa-bb97-5ca79c6e244f service nova] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Received event network-vif-deleted-5f04dc27-a2d4-46cc-bf3c-947c4c3dab59 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.594112] env[62235]: DEBUG nova.compute.manager [req-770072bd-2bc7-4217-952b-b66385aa53ef req-183652e9-eb42-43fa-bb97-5ca79c6e244f service nova] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Received event network-vif-deleted-19984205-bfbe-4c4a-a3e6-389954e0aa5a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.598735] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.736414] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.736836] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.737943] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda6063d-6ffc-459f-bcca-541ad0e91560 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.763741] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.764934] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c36ae10-f61e-4781-9b19-df6cd6cf50ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.797351] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfiguring VM to detach interface {{(pid=62235) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 879.799205] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f4a8b9c-a006-4a69-b32f-410857f6af8e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.817286] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.817552] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.827100] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 879.827100] env[62235]: value = "task-1271819" [ 879.827100] env[62235]: _type = "Task" [ 879.827100] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.843760] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.901418] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.901775] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.938858] env[62235]: ERROR nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [req-e348fed1-dbf8-4fa4-8ba8-2aa1d0fb4107] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e348fed1-dbf8-4fa4-8ba8-2aa1d0fb4107"}]} [ 879.958844] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 879.976953] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 879.977120] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 879.989737] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 880.010131] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 880.203053] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bb234a-94f4-4938-a080-4b5f37c28055 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.212225] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12056c03-bc7d-4861-a994-3e0163fa26bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.248079] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4409302c-df29-4603-8d13-c9c2e148ae65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.257795] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab48c2e4-1603-4da7-96db-be3bb0495acf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.273863] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 880.320693] env[62235]: DEBUG nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.338236] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.405573] env[62235]: DEBUG nova.compute.utils [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.805509] env[62235]: DEBUG nova.scheduler.client.report [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updated inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 880.805801] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 91 to 92 during operation: update_inventory {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 880.805988] env[62235]: DEBUG nova.compute.provider_tree [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 880.841426] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.845068] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.908613] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.312287] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.948s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.315517] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.707s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.315825] env[62235]: DEBUG nova.objects.instance [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lazy-loading 'resources' on Instance uuid d165f549-1759-41bf-86fd-b77793bd49c0 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.342254] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.825097] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b47b19fb-2064-4446-bf11-c6350afd8762 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.628s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.827584] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.227s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.827584] env[62235]: INFO nova.compute.manager [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Unshelving [ 881.843484] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.981527] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.981776] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.982008] env[62235]: INFO nova.compute.manager [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Attaching volume 4cc49965-d1d4-4f4d-84ed-8309d42ebfb0 to /dev/sdc [ 882.017554] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce068d8-5e86-4bf3-95ed-07791b7937ff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.023518] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2867e47c-8594-4670-b7a7-9972984493c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.036138] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521b21a1-0208-425f-9398-88c1aa536aa2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.041605] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d079580-8bff-4bb9-b8ff-5ef549c1f88e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.083853] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fe444a-f9e6-41d0-9c11-0d85a02ceb77 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.086726] env[62235]: DEBUG nova.virt.block_device [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updating existing volume attachment record: 90fb219f-af37-4f18-88ee-0dd18abe3748 {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 882.095278] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605725cd-9ad3-4ad7-aebe-d39717490c47 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.110798] env[62235]: DEBUG nova.compute.provider_tree [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.344845] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.614233] env[62235]: DEBUG nova.scheduler.client.report [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.841126] env[62235]: DEBUG nova.compute.utils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.849604] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.120711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.805s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.123206] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.396s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.144433] env[62235]: INFO nova.scheduler.client.report [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Deleted allocations for instance d165f549-1759-41bf-86fd-b77793bd49c0 [ 883.348767] env[62235]: INFO nova.virt.block_device [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Booting with volume ff64a51b-f62a-4682-83fb-4e26eae16346 at /dev/sdb [ 883.350261] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.382569] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16a19964-94fc-4829-9bb9-16313dad252e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.393215] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc70ccb-6407-47d3-8598-ba01475a3c05 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.425104] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0ee1203-b5af-494c-aa2c-784cc2076190 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.435171] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4294f974-c8b4-403d-8c06-01b28671b6ee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.465359] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ebde67-bc0f-47a1-bde2-8579505c7c96 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.473517] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c1e4f6-4b5a-4cdf-bbd3-d304bf59acd1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.489546] env[62235]: DEBUG nova.virt.block_device [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating existing volume attachment record: a4d7f10b-b6f8-4e4b-bc40-2ac72b5d7820 {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 883.629621] env[62235]: INFO nova.compute.claims [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.657676] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4190446-ee18-4917-84e3-e3e63c932d68 tempest-FloatingIPsAssociationTestJSON-1983203801 tempest-FloatingIPsAssociationTestJSON-1983203801-project-member] Lock "d165f549-1759-41bf-86fd-b77793bd49c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.054s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.847987] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.137088] env[62235]: INFO nova.compute.resource_tracker [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating resource usage from migration 18afe82b-d5ea-43a2-9aad-5add4f0c891b [ 884.350029] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.375955] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2711d5-3b4d-42a0-a62d-f90b06adfb78 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.387355] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5dbc37-81e7-4448-b921-47d70aaef26e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.427609] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4e4ece-adc3-4de0-a438-4c5235502c0a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.436839] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b77cf79-e3d4-4661-8ad0-0cd3aad19d0f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.454391] env[62235]: DEBUG nova.compute.provider_tree [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.849131] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.958041] env[62235]: DEBUG nova.scheduler.client.report [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.063579] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.063929] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.064265] env[62235]: INFO nova.compute.manager [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Rebooting instance [ 885.350343] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.464792] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.342s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.465022] env[62235]: INFO nova.compute.manager [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Migrating [ 885.465366] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.465551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.467040] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.375s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.468714] env[62235]: DEBUG nova.objects.instance [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lazy-loading 'resources' on Instance uuid 80a803d1-c034-42b6-a15d-9fc3893953ad {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.798441] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c40dae-a5ed-cacb-db5d-0dae37bf58f7/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 885.799437] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced72b69-68a3-4d77-8c43-d437901ace0d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.806237] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c40dae-a5ed-cacb-db5d-0dae37bf58f7/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 885.806444] env[62235]: ERROR oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c40dae-a5ed-cacb-db5d-0dae37bf58f7/disk-0.vmdk due to incomplete transfer. [ 885.806676] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-008c211a-5575-486f-94e8-774625cb6849 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.813696] env[62235]: DEBUG oslo_vmware.rw_handles [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c40dae-a5ed-cacb-db5d-0dae37bf58f7/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 885.813901] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Uploaded image b1ebb36c-537b-4e31-9dfe-858f1f81662b to the Glance image server {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 885.816282] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 885.816586] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-91001a94-8e20-42e8-b0cf-9ec6db642ff6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.823125] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 885.823125] env[62235]: value = "task-1271825" [ 885.823125] env[62235]: _type = "Task" [ 885.823125] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.831553] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271825, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.849313] env[62235]: DEBUG oslo_vmware.api [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271819, 'name': ReconfigVM_Task, 'duration_secs': 5.819101} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.849607] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.849827] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Reconfigured VM to detach interface {{(pid=62235) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 885.970935] env[62235]: INFO nova.compute.rpcapi [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 885.971535] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.145426] env[62235]: DEBUG nova.compute.manager [req-37f80408-8191-41e0-9eb9-9c40c8851256 req-8a499908-4fac-42b4-9b44-b92bfd37187f service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-vif-deleted-c7098635-f26b-4913-9096-7241e46f0a61 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.145704] env[62235]: INFO nova.compute.manager [req-37f80408-8191-41e0-9eb9-9c40c8851256 req-8a499908-4fac-42b4-9b44-b92bfd37187f service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Neutron deleted interface c7098635-f26b-4913-9096-7241e46f0a61; detaching it from the instance and deleting it from the info cache [ 886.146012] env[62235]: DEBUG nova.network.neutron [req-37f80408-8191-41e0-9eb9-9c40c8851256 req-8a499908-4fac-42b4-9b44-b92bfd37187f service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f80efd23-9c8c-4c33-8d50-e2bbbb4f8770", "address": "fa:16:3e:1c:4d:69", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf80efd23-9c", "ovs_interfaceid": "f80efd23-9c8c-4c33-8d50-e2bbbb4f8770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.177899] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8fc913-85d4-4316-823f-3c4b9d8d9003 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.186751] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca7e74d-4f84-4485-9a0b-61deb37d88f5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.217600] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836eb61c-886e-4f57-b846-474d295dd455 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.225950] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8018f9a-7460-48ea-ac67-18e312ae873f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.239751] env[62235]: DEBUG nova.compute.provider_tree [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.336735] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271825, 'name': Destroy_Task, 'duration_secs': 0.315974} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.336857] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Destroyed the VM [ 886.337551] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 886.339156] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-451076d3-1c30-4590-b382-68864c758f63 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.348090] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 886.348090] env[62235]: value = "task-1271827" [ 886.348090] env[62235]: _type = "Task" [ 886.348090] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.359234] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271827, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.491564] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.491760] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.491947] env[62235]: DEBUG nova.network.neutron [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.633832] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 886.634180] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273465', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'name': 'volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'serial': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 886.635283] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01b368a-394a-4d18-9b30-f1cb4a409ecf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.654253] env[62235]: DEBUG oslo_concurrency.lockutils [req-37f80408-8191-41e0-9eb9-9c40c8851256 req-8a499908-4fac-42b4-9b44-b92bfd37187f service nova] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.654494] env[62235]: DEBUG oslo_concurrency.lockutils [req-37f80408-8191-41e0-9eb9-9c40c8851256 req-8a499908-4fac-42b4-9b44-b92bfd37187f service nova] Acquired lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.655664] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa59414-d55b-46c5-83c4-60d3766cd56d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.659054] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8d1b5b-a5e5-4a01-8ed2-6b1c9a1d6bf1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.662499] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.662961] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.663205] env[62235]: DEBUG nova.network.neutron [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.683723] env[62235]: DEBUG oslo_concurrency.lockutils [req-37f80408-8191-41e0-9eb9-9c40c8851256 req-8a499908-4fac-42b4-9b44-b92bfd37187f service nova] Releasing lock "ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.684235] env[62235]: WARNING nova.compute.manager [req-37f80408-8191-41e0-9eb9-9c40c8851256 req-8a499908-4fac-42b4-9b44-b92bfd37187f service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Detach interface failed, port_id=c7098635-f26b-4913-9096-7241e46f0a61, reason: No device with interface-id c7098635-f26b-4913-9096-7241e46f0a61 exists on VM: nova.exception.NotFound: No device with interface-id c7098635-f26b-4913-9096-7241e46f0a61 exists on VM [ 886.715880] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0/volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.716647] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6ec348a-2b57-4d1a-9cdc-c43521896683 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.739163] env[62235]: DEBUG oslo_vmware.api [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 886.739163] env[62235]: value = "task-1271828" [ 886.739163] env[62235]: _type = "Task" [ 886.739163] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.743631] env[62235]: DEBUG nova.scheduler.client.report [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.753853] env[62235]: DEBUG oslo_vmware.api [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271828, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.859453] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271827, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.122853] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.123091] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.123542] env[62235]: DEBUG nova.network.neutron [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.249653] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.251652] env[62235]: DEBUG oslo_vmware.api [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271828, 'name': ReconfigVM_Task, 'duration_secs': 0.40694} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.252172] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.447s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.253647] env[62235]: INFO nova.compute.claims [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.256597] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfigured VM instance instance-00000042 to attach disk [datastore1] volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0/volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.262104] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7cf032b-e779-40ae-87a3-62080805324c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.274062] env[62235]: INFO nova.scheduler.client.report [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Deleted allocations for instance 80a803d1-c034-42b6-a15d-9fc3893953ad [ 887.282320] env[62235]: DEBUG oslo_vmware.api [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 887.282320] env[62235]: value = "task-1271829" [ 887.282320] env[62235]: _type = "Task" [ 887.282320] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.292438] env[62235]: DEBUG oslo_vmware.api [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271829, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.320185] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.320458] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "ced2aa67-ce75-4af5-9c55-773835d05733" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.320748] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.320945] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.321133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "ced2aa67-ce75-4af5-9c55-773835d05733-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.323499] env[62235]: DEBUG nova.network.neutron [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance_info_cache with network_info: [{"id": "3c9265cd-f975-407d-8f23-ab587b57e089", "address": "fa:16:3e:b3:9f:3c", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c9265cd-f9", "ovs_interfaceid": "3c9265cd-f975-407d-8f23-ab587b57e089", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.325383] env[62235]: INFO nova.compute.manager [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Terminating instance [ 887.327357] env[62235]: DEBUG nova.compute.manager [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.327556] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.328508] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc38385f-1c85-47f7-8989-034e6f93f537 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.338255] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.338575] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a9d3b1e-166e-473d-b541-9f2396973470 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.347557] env[62235]: DEBUG oslo_vmware.api [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 887.347557] env[62235]: value = "task-1271830" [ 887.347557] env[62235]: _type = "Task" [ 887.347557] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.360288] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271827, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.363840] env[62235]: DEBUG oslo_vmware.api [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.503248] env[62235]: DEBUG nova.network.neutron [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.787751] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b5aafa78-1c2e-437a-9526-8b92aa7e6c4f tempest-ServersV294TestFqdnHostnames-1722525471 tempest-ServersV294TestFqdnHostnames-1722525471-project-member] Lock "80a803d1-c034-42b6-a15d-9fc3893953ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.981s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.797271] env[62235]: DEBUG oslo_vmware.api [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271829, 'name': ReconfigVM_Task, 'duration_secs': 0.151481} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.800268] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273465', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'name': 'volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'serial': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 887.826313] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.860226] env[62235]: DEBUG oslo_vmware.api [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271830, 'name': PowerOffVM_Task, 'duration_secs': 0.24775} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.864206] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.864389] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.865043] env[62235]: DEBUG oslo_vmware.api [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271827, 'name': RemoveSnapshot_Task, 'duration_secs': 1.222093} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.865043] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16c9b402-5bc7-4d23-b2d0-3f62c3d21c2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.866311] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 887.866551] env[62235]: INFO nova.compute.manager [None req-4c7d2d6b-07c1-45c3-8c33-0731d39d4ab2 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Took 13.95 seconds to snapshot the instance on the hypervisor. [ 887.969415] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.969886] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.970234] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleting the datastore file [datastore2] ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.972737] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02068a6a-91b6-4113-aaee-03959e1a96bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.983022] env[62235]: DEBUG oslo_vmware.api [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 887.983022] env[62235]: value = "task-1271832" [ 887.983022] env[62235]: _type = "Task" [ 887.983022] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.989517] env[62235]: DEBUG oslo_vmware.api [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271832, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.990470] env[62235]: INFO nova.network.neutron [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Port f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 887.990820] env[62235]: DEBUG nova.network.neutron [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [{"id": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "address": "fa:16:3e:8b:e1:18", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap235e1a30-ec", "ovs_interfaceid": "235e1a30-ec5f-4d9a-8085-8994ac4c8314", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.009583] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.011847] env[62235]: DEBUG nova.compute.manager [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.012984] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f03203-d745-4799-b0c6-f7e069809591 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.182226] env[62235]: DEBUG nova.compute.manager [req-f7842e71-dfa6-4dc6-ab96-06a611b91dc4 req-4929bc0f-59d6-4d52-9310-e41d0adc800d service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-vif-deleted-f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.431164] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18a7bfb-a2f8-427c-8449-bd98df7fdd87 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.441089] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b3381a-28af-41dc-9e05-e0e77c7a8d48 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.472752] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb0d22a-975c-4fe1-872e-2e86eaa54461 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.483965] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de71ca57-4379-44a2-9b4a-439ae98e5570 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.493467] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-ced2aa67-ce75-4af5-9c55-773835d05733" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.495449] env[62235]: DEBUG oslo_vmware.api [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271832, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175229} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.503526] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.503725] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.503912] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.504105] env[62235]: INFO nova.compute.manager [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Took 1.18 seconds to destroy the instance on the hypervisor. [ 888.504355] env[62235]: DEBUG oslo.service.loopingcall [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.504765] env[62235]: DEBUG nova.compute.provider_tree [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.506291] env[62235]: DEBUG nova.compute.manager [-] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.506395] env[62235]: DEBUG nova.network.neutron [-] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 888.554153] env[62235]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 could not be found.", "detail": ""}} {{(pid=62235) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 888.554396] env[62235]: DEBUG nova.network.neutron [-] Unable to show port f80efd23-9c8c-4c33-8d50-e2bbbb4f8770 as it no longer exists. {{(pid=62235) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 888.839285] env[62235]: DEBUG nova.objects.instance [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lazy-loading 'flavor' on Instance uuid 06ec6b9c-bef0-4f96-a185-4315961de7f7 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.006195] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2412021c-6977-4fde-939a-66e2eb033996 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-ced2aa67-ce75-4af5-9c55-773835d05733-c7098635-f26b-4913-9096-7241e46f0a61" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.774s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.009670] env[62235]: DEBUG nova.scheduler.client.report [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.034175] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9904105-7698-458a-a03e-5322767d5868 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.043855] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Doing hard reboot of VM {{(pid=62235) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 889.045450] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-89058c01-bed2-4230-acc2-3930ff7eaec5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.056492] env[62235]: DEBUG oslo_vmware.api [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 889.056492] env[62235]: value = "task-1271833" [ 889.056492] env[62235]: _type = "Task" [ 889.056492] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.070095] env[62235]: DEBUG oslo_vmware.api [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271833, 'name': ResetVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.098325] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.351125] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea10a33-cca1-4e7d-abca-b6a1c6745f09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.357557] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5773e4f0-fdd3-4538-bf56-470f9d8cb342 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.376s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.377198] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance '44ed8739-ad20-41a3-abe1-38176c1bee25' progress to 0 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 889.469367] env[62235]: DEBUG nova.network.neutron [-] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.521738] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.521839] env[62235]: DEBUG nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.526973] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.378s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.529046] env[62235]: INFO nova.compute.claims [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.568318] env[62235]: DEBUG oslo_vmware.api [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271833, 'name': ResetVM_Task, 'duration_secs': 0.088957} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.568318] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Did hard reboot of VM {{(pid=62235) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 889.568318] env[62235]: DEBUG nova.compute.manager [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.569195] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7222cbe2-d92a-44a8-bfd9-511bf0399f8c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.734733] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.735048] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.883939] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.884283] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-056a94e5-861f-4e1d-ba08-e0aa099d49eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.893347] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 889.893347] env[62235]: value = "task-1271834" [ 889.893347] env[62235]: _type = "Task" [ 889.893347] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.904239] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271834, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.972546] env[62235]: INFO nova.compute.manager [-] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Took 1.47 seconds to deallocate network for instance. [ 890.028922] env[62235]: DEBUG nova.compute.utils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.030509] env[62235]: DEBUG nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 890.030792] env[62235]: DEBUG nova.network.neutron [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 890.091037] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9c5ae16f-b899-4f8d-a7df-534b3b59924e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.027s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.114954] env[62235]: DEBUG nova.policy [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96af7a114d04c0dad3c53f5a2dcd2bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef6547723ccf4be891da50a179dae06f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 890.217730] env[62235]: DEBUG nova.compute.manager [req-a8e46416-c37c-47e4-a8ee-b5634f9ae75e req-72c6a067-7f4f-4b47-8650-e791e584a362 service nova] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Received event network-vif-deleted-235e1a30-ec5f-4d9a-8085-8994ac4c8314 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.239399] env[62235]: INFO nova.compute.manager [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Detaching volume 3ee292e3-3b28-4582-8270-d1b850f04967 [ 890.285226] env[62235]: INFO nova.virt.block_device [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Attempting to driver detach volume 3ee292e3-3b28-4582-8270-d1b850f04967 from mountpoint /dev/sdb [ 890.286292] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 890.286598] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273462', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'name': 'volume-3ee292e3-3b28-4582-8270-d1b850f04967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'serial': '3ee292e3-3b28-4582-8270-d1b850f04967'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 890.288980] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfa7662-2745-44be-952b-1f5e69c7a2fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.314983] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea4169f-a00f-4c99-80e4-76c4219d0a62 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.323302] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e199a2-2c00-4fa2-985b-445856dc3f42 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.349628] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123b9f65-c067-4eb6-88e8-11118d59552d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.366536] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] The volume has not been displaced from its original location: [datastore1] volume-3ee292e3-3b28-4582-8270-d1b850f04967/volume-3ee292e3-3b28-4582-8270-d1b850f04967.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 890.372734] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 890.373106] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-849f1c91-ad9a-46de-9a6a-1416db0e3a0e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.395564] env[62235]: DEBUG oslo_vmware.api [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 890.395564] env[62235]: value = "task-1271835" [ 890.395564] env[62235]: _type = "Task" [ 890.395564] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.407724] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271834, 'name': PowerOffVM_Task, 'duration_secs': 0.189154} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.410624] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.410837] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance '44ed8739-ad20-41a3-abe1-38176c1bee25' progress to 17 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 890.415101] env[62235]: DEBUG oslo_vmware.api [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271835, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.415340] env[62235]: DEBUG nova.network.neutron [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Successfully created port: 50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.480822] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.534380] env[62235]: DEBUG nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.677941] env[62235]: DEBUG nova.compute.manager [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.678927] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-761c3f8b-9d81-4e54-bf4f-98cb39548cfd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.721356] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d5e409-d772-4138-b616-3dd65a7981a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.729318] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a60ec09-7170-4846-aa1a-a919849516fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.765037] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0799bb2-9a25-45ec-a233-b8607534f695 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.772686] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d2c416-4e21-4e7c-8a5f-1849f51cbde0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.788900] env[62235]: DEBUG nova.compute.provider_tree [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.909923] env[62235]: DEBUG oslo_vmware.api [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271835, 'name': ReconfigVM_Task, 'duration_secs': 0.265236} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.910487] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 890.915729] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8eba15a1-dc03-49c4-8ac6-28185ff9dbf8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.928690] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.932014] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.939283] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39aab76b-a481-40af-bc66-c51b2c037435 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.956689] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 890.956689] env[62235]: value = "task-1271836" [ 890.956689] env[62235]: _type = "Task" [ 890.956689] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.963038] env[62235]: DEBUG oslo_vmware.api [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 890.963038] env[62235]: value = "task-1271837" [ 890.963038] env[62235]: _type = "Task" [ 890.963038] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.977839] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271836, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.978480] env[62235]: DEBUG oslo_vmware.api [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271837, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.193757] env[62235]: INFO nova.compute.manager [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] instance snapshotting [ 891.197605] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3036ff-4006-42ce-ae04-a5824ce39dcb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.215434] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8618a980-a49e-4ec4-9cb8-bd5f63123115 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.291471] env[62235]: DEBUG nova.scheduler.client.report [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.472083] env[62235]: DEBUG oslo_vmware.api [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.475090] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271836, 'name': ReconfigVM_Task, 'duration_secs': 0.287542} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.475419] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance '44ed8739-ad20-41a3-abe1-38176c1bee25' progress to 33 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 891.547789] env[62235]: DEBUG nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.574513] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.574835] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.575061] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.575301] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.575488] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.575674] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.575992] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.576221] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.576496] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.576644] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.576856] env[62235]: DEBUG nova.virt.hardware [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.577772] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ecc636-372f-4af6-aafc-41dc240a49e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.586897] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c947c77b-ded2-4586-ac59-87e7d42d3422 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.322951] env[62235]: DEBUG nova.network.neutron [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Successfully updated port: 50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.326067] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 892.326879] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.800s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.327237] env[62235]: DEBUG nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.331768] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 892.332059] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 892.332293] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.332490] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 892.332767] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.332828] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 892.333033] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 892.333315] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 892.333461] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 892.333695] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 892.333891] env[62235]: DEBUG nova.virt.hardware [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.341507] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Reconfiguring VM instance instance-0000004c to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 892.342233] env[62235]: DEBUG nova.compute.manager [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Received event network-vif-plugged-50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.342441] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] Acquiring lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.342656] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.342831] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.342999] env[62235]: DEBUG nova.compute.manager [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] No waiting events found dispatching network-vif-plugged-50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.343191] env[62235]: WARNING nova.compute.manager [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Received unexpected event network-vif-plugged-50f22daa-f003-4786-a006-57fc3b25326d for instance with vm_state building and task_state spawning. [ 892.343345] env[62235]: DEBUG nova.compute.manager [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Received event network-changed-50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.343497] env[62235]: DEBUG nova.compute.manager [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Refreshing instance network info cache due to event network-changed-50f22daa-f003-4786-a006-57fc3b25326d. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 892.343679] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] Acquiring lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.343820] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] Acquired lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.343975] env[62235]: DEBUG nova.network.neutron [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Refreshing network info cache for port 50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 892.345837] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.346627] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f02aa87f-9f8c-41ce-8b10-526f60e0ad7d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.348726] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.585s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.348968] env[62235]: DEBUG nova.objects.instance [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lazy-loading 'resources' on Instance uuid 63e16fd4-3a5a-4231-9013-c6a01f0eab80 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.353072] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3d197fc-da44-4eb1-bdca-ec0f0d2699c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.386462] env[62235]: DEBUG oslo_vmware.api [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271837, 'name': ReconfigVM_Task, 'duration_secs': 0.775791} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.388808] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273462', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'name': 'volume-3ee292e3-3b28-4582-8270-d1b850f04967', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '3ee292e3-3b28-4582-8270-d1b850f04967', 'serial': '3ee292e3-3b28-4582-8270-d1b850f04967'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 892.391655] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 892.391655] env[62235]: value = "task-1271838" [ 892.391655] env[62235]: _type = "Task" [ 892.391655] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.392378] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 892.392378] env[62235]: value = "task-1271839" [ 892.392378] env[62235]: _type = "Task" [ 892.392378] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.405947] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271839, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.410126] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271838, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.515541] env[62235]: DEBUG nova.network.neutron [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.598354] env[62235]: DEBUG nova.network.neutron [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.848380] env[62235]: DEBUG nova.compute.utils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.850007] env[62235]: DEBUG nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.850198] env[62235]: DEBUG nova.network.neutron [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 892.905877] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271838, 'name': ReconfigVM_Task, 'duration_secs': 0.227657} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.910018] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Reconfigured VM instance instance-0000004c to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 892.910018] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271839, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.912325] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7f4904-c471-4e6a-83db-bce8ded9e12c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.943403] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 44ed8739-ad20-41a3-abe1-38176c1bee25/44ed8739-ad20-41a3-abe1-38176c1bee25.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.946739] env[62235]: DEBUG nova.policy [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d7ef1bc26534632ad28710f9df9803c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e22686dbba4420dad98d5de1f4fd449', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.951507] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc8f1f30-edc3-4f68-bfad-979c7569fcd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.977549] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 892.977549] env[62235]: value = "task-1271840" [ 892.977549] env[62235]: _type = "Task" [ 892.977549] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.988546] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271840, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.994541] env[62235]: DEBUG nova.objects.instance [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lazy-loading 'flavor' on Instance uuid 06ec6b9c-bef0-4f96-a185-4315961de7f7 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.103144] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a73f64-62e7-4940-967a-4a040b270255 req-62c53f68-74d7-4fdd-b70d-d0f164e08a3e service nova] Releasing lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.103144] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.103144] env[62235]: DEBUG nova.network.neutron [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.141706] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9905cc-fd6e-4355-aa9e-6d22903e0c29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.150623] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cac061-b0c3-4d76-b090-90bf4bccd49f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.187423] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae52297-22c1-48e9-a804-d20fa64198a9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.196373] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5dceb8d-efeb-48f6-a640-9ec83fb68bf2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.212820] env[62235]: DEBUG nova.compute.provider_tree [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.354013] env[62235]: DEBUG nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.357550] env[62235]: DEBUG nova.network.neutron [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Successfully created port: 347663dc-4c1e-4e45-ab26-82dc53ace862 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.405289] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271839, 'name': CreateSnapshot_Task, 'duration_secs': 0.743102} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.406594] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 893.409402] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b482a06a-95ac-4a29-ab04-778298423fd8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.493660] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271840, 'name': ReconfigVM_Task, 'duration_secs': 0.33287} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.493660] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 44ed8739-ad20-41a3-abe1-38176c1bee25/44ed8739-ad20-41a3-abe1-38176c1bee25.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.493660] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance '44ed8739-ad20-41a3-abe1-38176c1bee25' progress to 50 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 893.668641] env[62235]: DEBUG nova.network.neutron [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 893.681153] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "ecb104db-2172-4b09-867a-7a2813e03fa5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.681475] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.006s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.715722] env[62235]: DEBUG nova.scheduler.client.report [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.866723] env[62235]: DEBUG nova.network.neutron [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Updating instance_info_cache with network_info: [{"id": "50f22daa-f003-4786-a006-57fc3b25326d", "address": "fa:16:3e:c5:de:67", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50f22daa-f0", "ovs_interfaceid": "50f22daa-f003-4786-a006-57fc3b25326d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.908420] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.908674] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.927136] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 893.927838] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b80762a0-c7d7-43f4-b56c-7004ae39582d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.938527] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 893.938527] env[62235]: value = "task-1271841" [ 893.938527] env[62235]: _type = "Task" [ 893.938527] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.947836] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271841, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.001393] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c9a7f2df-b011-4be9-8cda-1c8e2e4dd9d6 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.266s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.003018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82221c8-4192-4709-8bbf-a658b62df6eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.025172] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d15601-2bfd-4dab-9db5-05f0520e1b81 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.046794] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance '44ed8739-ad20-41a3-abe1-38176c1bee25' progress to 67 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 894.097533] env[62235]: DEBUG oslo_concurrency.lockutils [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.097842] env[62235]: DEBUG oslo_concurrency.lockutils [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.187209] env[62235]: DEBUG nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.223019] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.872s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.224249] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.383s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.226135] env[62235]: INFO nova.compute.claims [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.247736] env[62235]: INFO nova.scheduler.client.report [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted allocations for instance 63e16fd4-3a5a-4231-9013-c6a01f0eab80 [ 894.369131] env[62235]: DEBUG nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.369937] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.370367] env[62235]: DEBUG nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Instance network_info: |[{"id": "50f22daa-f003-4786-a006-57fc3b25326d", "address": "fa:16:3e:c5:de:67", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50f22daa-f0", "ovs_interfaceid": "50f22daa-f003-4786-a006-57fc3b25326d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 894.371631] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:de:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '50f22daa-f003-4786-a006-57fc3b25326d', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.380538] env[62235]: DEBUG oslo.service.loopingcall [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.380897] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.381361] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7506f63-7925-4a40-8ff2-aa2c0f3bf3d5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.413793] env[62235]: DEBUG nova.compute.utils [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.416032] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.416032] env[62235]: value = "task-1271842" [ 894.416032] env[62235]: _type = "Task" [ 894.416032] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.418209] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.418446] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.418607] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.418845] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.419020] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.419782] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.419992] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.420182] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.420360] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.420596] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.420706] env[62235]: DEBUG nova.virt.hardware [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.421617] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a459aa7-743b-4b8a-ad47-b958a9a4f5c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.434562] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271842, 'name': CreateVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.438241] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9697c5-7ceb-4cf3-bc75-1626886e80db {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.451029] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271841, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.589522] env[62235]: DEBUG nova.network.neutron [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Port 3c9265cd-f975-407d-8f23-ab587b57e089 binding to destination host cpu-1 is already ACTIVE {{(pid=62235) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 894.601708] env[62235]: INFO nova.compute.manager [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Detaching volume 4cc49965-d1d4-4f4d-84ed-8309d42ebfb0 [ 894.640071] env[62235]: INFO nova.virt.block_device [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Attempting to driver detach volume 4cc49965-d1d4-4f4d-84ed-8309d42ebfb0 from mountpoint /dev/sdc [ 894.640338] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 894.640533] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273465', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'name': 'volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'serial': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 894.642110] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18435849-ba80-435a-9fa1-73fea412c5c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.666717] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b801bbc7-c663-46ac-9295-a173571867cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.675142] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7142dc-8f7e-4130-8497-5841ed60823a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.698855] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9904f20-c9da-49c8-86cd-678c6fb8a000 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.718538] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] The volume has not been displaced from its original location: [datastore1] volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0/volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 894.723769] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfiguring VM instance instance-00000042 to detach disk 2002 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 894.725084] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.725344] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03e63e11-3313-40a8-93ae-b56d7b55b70e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.746750] env[62235]: DEBUG oslo_vmware.api [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 894.746750] env[62235]: value = "task-1271843" [ 894.746750] env[62235]: _type = "Task" [ 894.746750] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.757940] env[62235]: DEBUG oslo_vmware.api [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271843, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.758968] env[62235]: DEBUG oslo_concurrency.lockutils [None req-589e2a77-2abd-482b-9a44-4e8af722d8d1 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "63e16fd4-3a5a-4231-9013-c6a01f0eab80" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.139s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.927075] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.017s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.933720] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271842, 'name': CreateVM_Task, 'duration_secs': 0.466555} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.933921] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.934588] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.934716] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.935324] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.935681] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c9d667a-187c-4c78-9b97-4f57d34561a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.942431] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 894.942431] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529468b0-1a8a-707f-ae84-13334a380be3" [ 894.942431] env[62235]: _type = "Task" [ 894.942431] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.957028] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271841, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.958522] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529468b0-1a8a-707f-ae84-13334a380be3, 'name': SearchDatastore_Task, 'duration_secs': 0.010698} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.958862] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.959131] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.959372] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.959525] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.959712] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.959994] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d36d34d3-8078-43dd-af53-c3022c78ec00 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.969628] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.969828] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 894.970625] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7231d7e5-0878-4bd2-a0ac-1da77d82016c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.978785] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 894.978785] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bef9e1-a3ec-393c-828a-d9784e49ee62" [ 894.978785] env[62235]: _type = "Task" [ 894.978785] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.987636] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bef9e1-a3ec-393c-828a-d9784e49ee62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.079988] env[62235]: DEBUG nova.network.neutron [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Successfully updated port: 347663dc-4c1e-4e45-ab26-82dc53ace862 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.149387] env[62235]: DEBUG nova.compute.manager [req-b4a84949-81f3-4e00-8a75-f93d92a74cda req-3225d1ef-ed41-47ae-8737-754e9388ce3f service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Received event network-vif-plugged-347663dc-4c1e-4e45-ab26-82dc53ace862 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.149618] env[62235]: DEBUG oslo_concurrency.lockutils [req-b4a84949-81f3-4e00-8a75-f93d92a74cda req-3225d1ef-ed41-47ae-8737-754e9388ce3f service nova] Acquiring lock "b5e4c6b2-4583-4a20-b570-99699063e172-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.149849] env[62235]: DEBUG oslo_concurrency.lockutils [req-b4a84949-81f3-4e00-8a75-f93d92a74cda req-3225d1ef-ed41-47ae-8737-754e9388ce3f service nova] Lock "b5e4c6b2-4583-4a20-b570-99699063e172-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.150305] env[62235]: DEBUG oslo_concurrency.lockutils [req-b4a84949-81f3-4e00-8a75-f93d92a74cda req-3225d1ef-ed41-47ae-8737-754e9388ce3f service nova] Lock "b5e4c6b2-4583-4a20-b570-99699063e172-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.150571] env[62235]: DEBUG nova.compute.manager [req-b4a84949-81f3-4e00-8a75-f93d92a74cda req-3225d1ef-ed41-47ae-8737-754e9388ce3f service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] No waiting events found dispatching network-vif-plugged-347663dc-4c1e-4e45-ab26-82dc53ace862 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.150759] env[62235]: WARNING nova.compute.manager [req-b4a84949-81f3-4e00-8a75-f93d92a74cda req-3225d1ef-ed41-47ae-8737-754e9388ce3f service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Received unexpected event network-vif-plugged-347663dc-4c1e-4e45-ab26-82dc53ace862 for instance with vm_state building and task_state spawning. [ 895.257035] env[62235]: DEBUG oslo_vmware.api [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271843, 'name': ReconfigVM_Task, 'duration_secs': 0.246078} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.257350] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Reconfigured VM instance instance-00000042 to detach disk 2002 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 895.264593] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9af603d-16c2-410f-92a7-60079749ae02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.283159] env[62235]: DEBUG oslo_vmware.api [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 895.283159] env[62235]: value = "task-1271844" [ 895.283159] env[62235]: _type = "Task" [ 895.283159] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.295593] env[62235]: DEBUG oslo_vmware.api [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271844, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.424128] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9c74be-f0ba-43fa-a0d7-8ac99fb04fdb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.432622] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0390b3-5a7a-4e27-8611-f629e83b5886 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.470134] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357a9c6e-3177-46e8-885f-24c240b241ec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.480529] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271841, 'name': CloneVM_Task, 'duration_secs': 1.504992} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.481935] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Created linked-clone VM from snapshot [ 895.485842] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac84092-9d85-4703-a4ce-af4cd3a59857 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.492097] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fa96f3-218a-4546-9cfe-960a58f4ca72 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.501112] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Uploading image 7e6294f3-c087-4ce3-bd45-09f6dccc5c97 {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 895.517625] env[62235]: DEBUG nova.compute.provider_tree [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.518895] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bef9e1-a3ec-393c-828a-d9784e49ee62, 'name': SearchDatastore_Task, 'duration_secs': 0.010209} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.522071] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5514708c-faea-49bf-b05c-5b4b094d38e4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.528923] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 895.528923] env[62235]: value = "vm-273469" [ 895.528923] env[62235]: _type = "VirtualMachine" [ 895.528923] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 895.529449] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4d7e652c-6630-4814-80b4-6cfd9f9fedcd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.532234] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 895.532234] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d08b7e-a0e8-7abd-860d-22241dd37268" [ 895.532234] env[62235]: _type = "Task" [ 895.532234] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.538584] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lease: (returnval){ [ 895.538584] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52becb62-b300-c420-f487-ff8c2d8b0626" [ 895.538584] env[62235]: _type = "HttpNfcLease" [ 895.538584] env[62235]: } obtained for exporting VM: (result){ [ 895.538584] env[62235]: value = "vm-273469" [ 895.538584] env[62235]: _type = "VirtualMachine" [ 895.538584] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 895.538921] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the lease: (returnval){ [ 895.538921] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52becb62-b300-c420-f487-ff8c2d8b0626" [ 895.538921] env[62235]: _type = "HttpNfcLease" [ 895.538921] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 895.542014] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d08b7e-a0e8-7abd-860d-22241dd37268, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.550981] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 895.550981] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52becb62-b300-c420-f487-ff8c2d8b0626" [ 895.550981] env[62235]: _type = "HttpNfcLease" [ 895.550981] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 895.551271] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 895.551271] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52becb62-b300-c420-f487-ff8c2d8b0626" [ 895.551271] env[62235]: _type = "HttpNfcLease" [ 895.551271] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 895.551977] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85467e9e-27e8-44b4-a6e8-34c25cb161e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.558552] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52df9530-d6af-58e4-c81a-51066ce5190b/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 895.558729] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52df9530-d6af-58e4-c81a-51066ce5190b/disk-0.vmdk for reading. {{(pid=62235) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 895.615475] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "refresh_cache-b5e4c6b2-4583-4a20-b570-99699063e172" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.615627] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "refresh_cache-b5e4c6b2-4583-4a20-b570-99699063e172" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.615777] env[62235]: DEBUG nova.network.neutron [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.635482] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.635711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.635889] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.658197] env[62235]: DEBUG nova.network.neutron [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.674789] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-64ad1a70-8e54-476f-9321-a54b7218e96f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.793529] env[62235]: DEBUG oslo_vmware.api [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271844, 'name': ReconfigVM_Task, 'duration_secs': 0.173497} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.793859] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273465', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'name': 'volume-4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '06ec6b9c-bef0-4f96-a185-4315961de7f7', 'attached_at': '', 'detached_at': '', 'volume_id': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0', 'serial': '4cc49965-d1d4-4f4d-84ed-8309d42ebfb0'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 895.797791] env[62235]: DEBUG nova.network.neutron [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Updating instance_info_cache with network_info: [{"id": "347663dc-4c1e-4e45-ab26-82dc53ace862", "address": "fa:16:3e:e3:ea:17", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap347663dc-4c", "ovs_interfaceid": "347663dc-4c1e-4e45-ab26-82dc53ace862", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.995990] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.996431] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.996697] env[62235]: INFO nova.compute.manager [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Attaching volume 99abe792-777d-4235-82bc-9e547c16e586 to /dev/sdb [ 896.022144] env[62235]: DEBUG nova.scheduler.client.report [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.031260] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc929510-a14c-4223-8c98-ecf972d95de6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.046162] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d08b7e-a0e8-7abd-860d-22241dd37268, 'name': SearchDatastore_Task, 'duration_secs': 0.016792} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.047074] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30825ea7-e30e-40d8-9483-a9c4d4bdd424 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.050011] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.050377] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d98ab48f-c325-473a-8ee8-57ecd3b81c8b/d98ab48f-c325-473a-8ee8-57ecd3b81c8b.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.051015] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c975d023-2600-469c-a8fc-4310f251452a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.061063] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 896.061063] env[62235]: value = "task-1271846" [ 896.061063] env[62235]: _type = "Task" [ 896.061063] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.069485] env[62235]: DEBUG nova.virt.block_device [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updating existing volume attachment record: e7fe781a-66d9-4cd0-a575-99611a00caaf {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 896.078091] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271846, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.305980] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "refresh_cache-b5e4c6b2-4583-4a20-b570-99699063e172" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.305980] env[62235]: DEBUG nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Instance network_info: |[{"id": "347663dc-4c1e-4e45-ab26-82dc53ace862", "address": "fa:16:3e:e3:ea:17", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap347663dc-4c", "ovs_interfaceid": "347663dc-4c1e-4e45-ab26-82dc53ace862", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.305980] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:ea:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '347663dc-4c1e-4e45-ab26-82dc53ace862', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.312270] env[62235]: DEBUG oslo.service.loopingcall [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.312796] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.313182] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2edc1f36-52ef-4812-81ea-e231801b8591 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.336530] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.336530] env[62235]: value = "task-1271850" [ 896.336530] env[62235]: _type = "Task" [ 896.336530] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.347018] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271850, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.349090] env[62235]: DEBUG nova.objects.instance [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lazy-loading 'flavor' on Instance uuid 06ec6b9c-bef0-4f96-a185-4315961de7f7 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.527774] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.529157] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.431s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.529462] env[62235]: DEBUG nova.objects.instance [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'pci_requests' on Instance uuid 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.572559] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271846, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491956} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.573106] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] d98ab48f-c325-473a-8ee8-57ecd3b81c8b/d98ab48f-c325-473a-8ee8-57ecd3b81c8b.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.573882] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.573882] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f95719e8-1173-4bbd-a502-9850a277b15a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.581996] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 896.581996] env[62235]: value = "task-1271851" [ 896.581996] env[62235]: _type = "Task" [ 896.581996] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.593478] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271851, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.707221] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.709957] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.709957] env[62235]: DEBUG nova.network.neutron [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.847753] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271850, 'name': CreateVM_Task, 'duration_secs': 0.475814} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.848141] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.848712] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.849019] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.849520] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.850116] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9e74cef-3dee-4f91-89c0-8b209fdd486d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.856953] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 896.856953] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522398ea-808e-9e29-0e6b-a91e132a2e54" [ 896.856953] env[62235]: _type = "Task" [ 896.856953] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.865868] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522398ea-808e-9e29-0e6b-a91e132a2e54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.034070] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "b0f0a771-529f-462b-a6a8-f4a8f86e9294" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.034070] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "b0f0a771-529f-462b-a6a8-f4a8f86e9294" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.036939] env[62235]: DEBUG nova.objects.instance [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'numa_topology' on Instance uuid 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.093661] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271851, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071203} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.094343] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 897.095277] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f703f4a-c1c0-42ba-bca3-c13aa4768bc4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.120629] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] d98ab48f-c325-473a-8ee8-57ecd3b81c8b/d98ab48f-c325-473a-8ee8-57ecd3b81c8b.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.120834] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0822c9db-a44c-467f-82f5-5a9609fb3b14 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.144058] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 897.144058] env[62235]: value = "task-1271852" [ 897.144058] env[62235]: _type = "Task" [ 897.144058] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.155674] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271852, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.178917] env[62235]: DEBUG nova.compute.manager [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Received event network-changed-347663dc-4c1e-4e45-ab26-82dc53ace862 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.179828] env[62235]: DEBUG nova.compute.manager [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Refreshing instance network info cache due to event network-changed-347663dc-4c1e-4e45-ab26-82dc53ace862. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.179828] env[62235]: DEBUG oslo_concurrency.lockutils [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] Acquiring lock "refresh_cache-b5e4c6b2-4583-4a20-b570-99699063e172" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.179828] env[62235]: DEBUG oslo_concurrency.lockutils [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] Acquired lock "refresh_cache-b5e4c6b2-4583-4a20-b570-99699063e172" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.179962] env[62235]: DEBUG nova.network.neutron [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Refreshing network info cache for port 347663dc-4c1e-4e45-ab26-82dc53ace862 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.356723] env[62235]: DEBUG oslo_concurrency.lockutils [None req-101e1888-62c3-4082-a3c9-a95b68df6624 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.259s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.370621] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522398ea-808e-9e29-0e6b-a91e132a2e54, 'name': SearchDatastore_Task, 'duration_secs': 0.009409} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.370621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.370621] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.370621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.370621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.370621] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.371081] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1469d48-7d7b-4569-b5c7-b94f62863891 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.382527] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.382527] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 897.382769] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e78c4bc-e03f-402b-85da-56fd190d13de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.393121] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 897.393121] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5270ac00-e820-635b-440d-6742eb58c301" [ 897.393121] env[62235]: _type = "Task" [ 897.393121] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.402103] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5270ac00-e820-635b-440d-6742eb58c301, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.467239] env[62235]: DEBUG nova.network.neutron [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance_info_cache with network_info: [{"id": "3c9265cd-f975-407d-8f23-ab587b57e089", "address": "fa:16:3e:b3:9f:3c", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c9265cd-f9", "ovs_interfaceid": "3c9265cd-f975-407d-8f23-ab587b57e089", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.535391] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "b0f0a771-529f-462b-a6a8-f4a8f86e9294" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.536402] env[62235]: DEBUG nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.539149] env[62235]: INFO nova.compute.claims [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.655804] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.907394] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5270ac00-e820-635b-440d-6742eb58c301, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.908435] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f570658-7864-4ebc-9310-f9b81be4a463 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.915646] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 897.915646] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5227590d-e878-f1e8-f20c-69b7e885cd84" [ 897.915646] env[62235]: _type = "Task" [ 897.915646] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.924416] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5227590d-e878-f1e8-f20c-69b7e885cd84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.937138] env[62235]: DEBUG nova.network.neutron [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Updated VIF entry in instance network info cache for port 347663dc-4c1e-4e45-ab26-82dc53ace862. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.937526] env[62235]: DEBUG nova.network.neutron [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Updating instance_info_cache with network_info: [{"id": "347663dc-4c1e-4e45-ab26-82dc53ace862", "address": "fa:16:3e:e3:ea:17", "network": {"id": "dbc89eb4-909f-4a62-af65-d3cfdf58b1ea", "bridge": "br-int", "label": "tempest-ImagesTestJSON-8388907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e22686dbba4420dad98d5de1f4fd449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap347663dc-4c", "ovs_interfaceid": "347663dc-4c1e-4e45-ab26-82dc53ace862", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.970180] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.044052] env[62235]: DEBUG nova.compute.utils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.047445] env[62235]: DEBUG nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.048628] env[62235]: DEBUG nova.network.neutron [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 898.096653] env[62235]: DEBUG nova.policy [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2cb98f6c8f548238697307ca1c29d34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a3e63e0ba8348ab917d44a623f8d117', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.155741] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271852, 'name': ReconfigVM_Task, 'duration_secs': 0.736548} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.155841] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Reconfigured VM instance instance-0000004e to attach disk [datastore1] d98ab48f-c325-473a-8ee8-57ecd3b81c8b/d98ab48f-c325-473a-8ee8-57ecd3b81c8b.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.156941] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-503540d0-93c4-4fdd-ae0e-017bec129540 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.165477] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 898.165477] env[62235]: value = "task-1271853" [ 898.165477] env[62235]: _type = "Task" [ 898.165477] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.180131] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271853, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.414106] env[62235]: DEBUG nova.network.neutron [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Successfully created port: 62b6b792-dba7-4eaf-b0d7-715fc1e84270 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.427267] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5227590d-e878-f1e8-f20c-69b7e885cd84, 'name': SearchDatastore_Task, 'duration_secs': 0.009902} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.427604] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.427878] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b5e4c6b2-4583-4a20-b570-99699063e172/b5e4c6b2-4583-4a20-b570-99699063e172.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.428177] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d992d4b-8d28-4adc-b517-70831a4f108c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.438470] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 898.438470] env[62235]: value = "task-1271855" [ 898.438470] env[62235]: _type = "Task" [ 898.438470] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.442132] env[62235]: DEBUG oslo_concurrency.lockutils [req-a93b47e7-5ec0-4761-a198-27c68398bb54 req-61e0ebb3-3564-4ec9-bcb4-0ee08fba97f3 service nova] Releasing lock "refresh_cache-b5e4c6b2-4583-4a20-b570-99699063e172" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.448345] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.497036] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09a999a-c67a-4ac5-8776-3b325abee9fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.518746] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0c5f9a-2668-4d5d-b471-cd079dd9c9e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.527821] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance '44ed8739-ad20-41a3-abe1-38176c1bee25' progress to 83 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 898.535609] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.535932] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.536210] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "06ec6b9c-bef0-4f96-a185-4315961de7f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.536523] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.536743] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.539012] env[62235]: INFO nova.compute.manager [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Terminating instance [ 898.540926] env[62235]: DEBUG nova.compute.manager [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.541169] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.542014] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec327d76-e7f3-4608-8539-3b714f1a3bbb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.547906] env[62235]: DEBUG nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.555753] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 898.556488] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65de936f-811d-4d90-8420-d750ac912273 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.565390] env[62235]: DEBUG oslo_vmware.api [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 898.565390] env[62235]: value = "task-1271856" [ 898.565390] env[62235]: _type = "Task" [ 898.565390] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.577410] env[62235]: DEBUG oslo_vmware.api [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.681742] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271853, 'name': Rename_Task, 'duration_secs': 0.163603} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.684990] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.685816] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7c271c5-9810-40ae-84e4-d8ce7db2c02f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.695551] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 898.695551] env[62235]: value = "task-1271857" [ 898.695551] env[62235]: _type = "Task" [ 898.695551] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.712703] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271857, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.782260] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e05f12-7104-4e3b-bf72-c833d01829c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.792189] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb1c8d7-d5b2-4fbe-9420-64bd6d83c2db {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.830646] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f0e7e4-1442-4781-ac8c-8c96a4105025 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.840490] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329eca77-2374-41a5-9d9a-7441124b3d4c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.859230] env[62235]: DEBUG nova.compute.provider_tree [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.949565] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464589} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.950318] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b5e4c6b2-4583-4a20-b570-99699063e172/b5e4c6b2-4583-4a20-b570-99699063e172.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.950593] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.950846] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e92ebf10-51fc-4b4d-92d8-c5b10a7708ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.960907] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 898.960907] env[62235]: value = "task-1271858" [ 898.960907] env[62235]: _type = "Task" [ 898.960907] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.971858] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271858, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.035019] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.035019] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20cd3f0a-33cb-449b-8551-81489e7f1dfe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.043885] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 899.043885] env[62235]: value = "task-1271859" [ 899.043885] env[62235]: _type = "Task" [ 899.043885] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.052696] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.077390] env[62235]: DEBUG oslo_vmware.api [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271856, 'name': PowerOffVM_Task, 'duration_secs': 0.212605} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.077977] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.077977] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 899.078310] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d08a350c-328f-45e7-bbef-6d3a99b16ecb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.155724] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 899.156021] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 899.156229] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Deleting the datastore file [datastore1] 06ec6b9c-bef0-4f96-a185-4315961de7f7 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.156534] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-553bce03-2417-481f-8d77-729f9fddbe0b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.164589] env[62235]: DEBUG oslo_vmware.api [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for the task: (returnval){ [ 899.164589] env[62235]: value = "task-1271861" [ 899.164589] env[62235]: _type = "Task" [ 899.164589] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.173401] env[62235]: DEBUG oslo_vmware.api [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271861, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.211679] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271857, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.362478] env[62235]: DEBUG nova.scheduler.client.report [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.473049] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070224} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.473049] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.473350] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84905202-e7b8-486f-aa95-e5819e89e054 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.497428] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] b5e4c6b2-4583-4a20-b570-99699063e172/b5e4c6b2-4583-4a20-b570-99699063e172.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.497762] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17c4edf7-774f-486c-b2d2-ca6dec523a39 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.522025] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 899.522025] env[62235]: value = "task-1271862" [ 899.522025] env[62235]: _type = "Task" [ 899.522025] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.527956] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271862, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.553149] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271859, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.562366] env[62235]: DEBUG nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.588417] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.588710] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.588878] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.589139] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.589701] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.589701] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.589701] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.590029] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.590130] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.590324] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.590548] env[62235]: DEBUG nova.virt.hardware [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.591473] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5ce12d-182d-4978-a40a-92ff96bed4d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.600850] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b5cbab-d218-482b-ae35-56e5998cebd7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.674810] env[62235]: DEBUG oslo_vmware.api [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Task: {'id': task-1271861, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142186} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.675125] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.675326] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.675506] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.675683] env[62235]: INFO nova.compute.manager [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 899.675938] env[62235]: DEBUG oslo.service.loopingcall [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.676145] env[62235]: DEBUG nova.compute.manager [-] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 899.676241] env[62235]: DEBUG nova.network.neutron [-] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 899.712516] env[62235]: DEBUG oslo_vmware.api [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271857, 'name': PowerOnVM_Task, 'duration_secs': 0.649175} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.712991] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.713250] env[62235]: INFO nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Took 8.17 seconds to spawn the instance on the hypervisor. [ 899.713509] env[62235]: DEBUG nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.714314] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3c0733-19c8-477e-8ac4-fb2a66de3408 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.868077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.339s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.870360] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.390s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.870679] env[62235]: DEBUG nova.objects.instance [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'resources' on Instance uuid ced2aa67-ce75-4af5-9c55-773835d05733 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.920868] env[62235]: INFO nova.network.neutron [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 900.030760] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271862, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.054707] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271859, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.189019] env[62235]: DEBUG nova.compute.manager [req-90c3b9f9-5556-4465-ade0-0be87bb7cfb3 req-57741f0e-ff62-437b-aff8-6e6ebd8b2378 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Received event network-vif-plugged-62b6b792-dba7-4eaf-b0d7-715fc1e84270 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.189019] env[62235]: DEBUG oslo_concurrency.lockutils [req-90c3b9f9-5556-4465-ade0-0be87bb7cfb3 req-57741f0e-ff62-437b-aff8-6e6ebd8b2378 service nova] Acquiring lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.189019] env[62235]: DEBUG oslo_concurrency.lockutils [req-90c3b9f9-5556-4465-ade0-0be87bb7cfb3 req-57741f0e-ff62-437b-aff8-6e6ebd8b2378 service nova] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.189019] env[62235]: DEBUG oslo_concurrency.lockutils [req-90c3b9f9-5556-4465-ade0-0be87bb7cfb3 req-57741f0e-ff62-437b-aff8-6e6ebd8b2378 service nova] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.189019] env[62235]: DEBUG nova.compute.manager [req-90c3b9f9-5556-4465-ade0-0be87bb7cfb3 req-57741f0e-ff62-437b-aff8-6e6ebd8b2378 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] No waiting events found dispatching network-vif-plugged-62b6b792-dba7-4eaf-b0d7-715fc1e84270 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.189019] env[62235]: WARNING nova.compute.manager [req-90c3b9f9-5556-4465-ade0-0be87bb7cfb3 req-57741f0e-ff62-437b-aff8-6e6ebd8b2378 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Received unexpected event network-vif-plugged-62b6b792-dba7-4eaf-b0d7-715fc1e84270 for instance with vm_state building and task_state spawning. [ 900.234798] env[62235]: INFO nova.compute.manager [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Took 21.45 seconds to build instance. [ 900.372619] env[62235]: DEBUG nova.network.neutron [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Successfully updated port: 62b6b792-dba7-4eaf-b0d7-715fc1e84270 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.533055] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271862, 'name': ReconfigVM_Task, 'duration_secs': 0.876476} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.533368] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Reconfigured VM instance instance-0000004f to attach disk [datastore1] b5e4c6b2-4583-4a20-b570-99699063e172/b5e4c6b2-4583-4a20-b570-99699063e172.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.534017] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d01eea4c-cba7-4ca2-88e2-bc5347832980 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.544368] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 900.544368] env[62235]: value = "task-1271863" [ 900.544368] env[62235]: _type = "Task" [ 900.544368] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.562173] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271863, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.568050] env[62235]: DEBUG oslo_vmware.api [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271859, 'name': PowerOnVM_Task, 'duration_secs': 1.400939} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.568724] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.568929] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a5715d-573a-43c3-96a6-ec1fc537686b tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance '44ed8739-ad20-41a3-abe1-38176c1bee25' progress to 100 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 900.580301] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1d5521-0c05-4d12-8c06-74d8eac778d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.590190] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1440e1af-db8d-46ae-a36a-6e4151dbc663 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.628532] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 900.628645] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273472', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'name': 'volume-99abe792-777d-4235-82bc-9e547c16e586', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15d41c0f-045d-4e1c-88b5-1841f99f2e74', 'attached_at': '', 'detached_at': '', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'serial': '99abe792-777d-4235-82bc-9e547c16e586'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 900.630378] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d715551-1107-4aee-9cdb-8e3599221672 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.634018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13cd62c4-c28c-4f83-a259-8bfdf15e627a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.654821] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72352ed0-04da-437f-a2d2-6f8032532f07 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.661036] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7b0638-0344-4a43-87ba-d14a4106958c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.676363] env[62235]: DEBUG nova.compute.provider_tree [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.701429] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] volume-99abe792-777d-4235-82bc-9e547c16e586/volume-99abe792-777d-4235-82bc-9e547c16e586.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.702078] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6716a3d8-c5a7-497d-a54b-430ee262f4e0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.727272] env[62235]: DEBUG oslo_vmware.api [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 900.727272] env[62235]: value = "task-1271864" [ 900.727272] env[62235]: _type = "Task" [ 900.727272] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.738484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e9551e66-e999-4183-8db7-b2ccbe2fa7d0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.961s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.738764] env[62235]: DEBUG oslo_vmware.api [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271864, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.885292] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "refresh_cache-3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.885510] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquired lock "refresh_cache-3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.885659] env[62235]: DEBUG nova.network.neutron [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.954248] env[62235]: DEBUG nova.network.neutron [-] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.060095] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271863, 'name': Rename_Task, 'duration_secs': 0.253117} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.060448] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.060712] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8dc65b71-703e-445c-9b85-2efe97f02c7a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.069170] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 901.069170] env[62235]: value = "task-1271865" [ 901.069170] env[62235]: _type = "Task" [ 901.069170] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.080807] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271865, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.179990] env[62235]: DEBUG nova.scheduler.client.report [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.220796] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.221301] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.221566] env[62235]: INFO nova.compute.manager [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Shelving [ 901.238347] env[62235]: DEBUG oslo_vmware.api [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271864, 'name': ReconfigVM_Task, 'duration_secs': 0.433663} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.238642] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Reconfigured VM instance instance-00000049 to attach disk [datastore1] volume-99abe792-777d-4235-82bc-9e547c16e586/volume-99abe792-777d-4235-82bc-9e547c16e586.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.244020] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b6dc9cf-d206-4ce2-ae31-635f4d664af6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.260775] env[62235]: DEBUG oslo_vmware.api [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 901.260775] env[62235]: value = "task-1271866" [ 901.260775] env[62235]: _type = "Task" [ 901.260775] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.275778] env[62235]: DEBUG oslo_vmware.api [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271866, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.420493] env[62235]: DEBUG nova.network.neutron [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 901.457784] env[62235]: INFO nova.compute.manager [-] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Took 1.78 seconds to deallocate network for instance. [ 901.583333] env[62235]: DEBUG nova.network.neutron [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Updating instance_info_cache with network_info: [{"id": "62b6b792-dba7-4eaf-b0d7-715fc1e84270", "address": "fa:16:3e:22:ad:a5", "network": {"id": "fc237f42-52cc-4043-8331-d796c72abb6c", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1222472131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e63e0ba8348ab917d44a623f8d117", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62b6b792-db", "ovs_interfaceid": "62b6b792-dba7-4eaf-b0d7-715fc1e84270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.586713] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271865, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.687046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.689837] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.965s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.691469] env[62235]: INFO nova.compute.claims [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.713845] env[62235]: INFO nova.scheduler.client.report [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleted allocations for instance ced2aa67-ce75-4af5-9c55-773835d05733 [ 901.729716] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.730128] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84a7c21a-d498-4732-a1cd-31565a630788 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.740042] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 901.740042] env[62235]: value = "task-1271867" [ 901.740042] env[62235]: _type = "Task" [ 901.740042] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.749312] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.775537] env[62235]: DEBUG oslo_vmware.api [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271866, 'name': ReconfigVM_Task, 'duration_secs': 0.144593} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.775942] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273472', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'name': 'volume-99abe792-777d-4235-82bc-9e547c16e586', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15d41c0f-045d-4e1c-88b5-1841f99f2e74', 'attached_at': '', 'detached_at': '', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'serial': '99abe792-777d-4235-82bc-9e547c16e586'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 901.968366] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.082727] env[62235]: DEBUG oslo_vmware.api [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271865, 'name': PowerOnVM_Task, 'duration_secs': 0.741309} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.083263] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.083346] env[62235]: INFO nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Took 7.72 seconds to spawn the instance on the hypervisor. [ 902.083527] env[62235]: DEBUG nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.084403] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4ac0af-6161-466c-bacd-3e32b12a3efd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.088503] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Releasing lock "refresh_cache-3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.088503] env[62235]: DEBUG nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Instance network_info: |[{"id": "62b6b792-dba7-4eaf-b0d7-715fc1e84270", "address": "fa:16:3e:22:ad:a5", "network": {"id": "fc237f42-52cc-4043-8331-d796c72abb6c", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1222472131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e63e0ba8348ab917d44a623f8d117", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62b6b792-db", "ovs_interfaceid": "62b6b792-dba7-4eaf-b0d7-715fc1e84270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.095807] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:ad:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c68b7663-4f0e-47f0-ac7f-40c6d952f7bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62b6b792-dba7-4eaf-b0d7-715fc1e84270', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.103944] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Creating folder: Project (5a3e63e0ba8348ab917d44a623f8d117). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 902.106518] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df589042-d8d5-40c7-bc37-b437f1a2b4cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.128076] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Created folder: Project (5a3e63e0ba8348ab917d44a623f8d117) in parent group-v273362. [ 902.128320] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Creating folder: Instances. Parent ref: group-v273474. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 902.129437] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1725af2-dbf8-4bbd-add7-c1b96391ca02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.144271] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Created folder: Instances in parent group-v273474. [ 902.144589] env[62235]: DEBUG oslo.service.loopingcall [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.144841] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 902.145126] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a90e837-9f53-4004-a979-a72a9cf5d904 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.168023] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.168023] env[62235]: value = "task-1271870" [ 902.168023] env[62235]: _type = "Task" [ 902.168023] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.177450] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271870, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.222425] env[62235]: DEBUG oslo_concurrency.lockutils [None req-53d3d0ac-9c09-40d9-b9b1-c2ab935566ea tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "ced2aa67-ce75-4af5-9c55-773835d05733" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.902s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.252310] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.332824] env[62235]: DEBUG nova.compute.manager [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Received event network-changed-62b6b792-dba7-4eaf-b0d7-715fc1e84270 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.332824] env[62235]: DEBUG nova.compute.manager [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Refreshing instance network info cache due to event network-changed-62b6b792-dba7-4eaf-b0d7-715fc1e84270. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.333178] env[62235]: DEBUG oslo_concurrency.lockutils [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] Acquiring lock "refresh_cache-3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.333437] env[62235]: DEBUG oslo_concurrency.lockutils [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] Acquired lock "refresh_cache-3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.333750] env[62235]: DEBUG nova.network.neutron [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Refreshing network info cache for port 62b6b792-dba7-4eaf-b0d7-715fc1e84270 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.626409] env[62235]: INFO nova.compute.manager [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Took 23.50 seconds to build instance. [ 902.681608] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271870, 'name': CreateVM_Task, 'duration_secs': 0.486829} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.681900] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.683178] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.683528] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.684067] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.684428] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16f8db34-b19f-483a-9b1c-d9c71e663bf0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.689630] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "44ed8739-ad20-41a3-abe1-38176c1bee25" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.690213] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.690288] env[62235]: DEBUG nova.compute.manager [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Going to confirm migration 1 {{(pid=62235) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 902.693396] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 902.693396] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52990d64-39ef-2f86-d734-0036fe9dcaf0" [ 902.693396] env[62235]: _type = "Task" [ 902.693396] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.710694] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52990d64-39ef-2f86-d734-0036fe9dcaf0, 'name': SearchDatastore_Task, 'duration_secs': 0.015247} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.710796] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.710974] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.714011] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.714011] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.714011] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.714011] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea44c337-ebfe-496b-a771-a53435da67da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.722791] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.722791] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.724754] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3448fa0-8d34-4f67-84f7-348e3faf5b58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.732991] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 902.732991] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5263fc0d-7bd9-2e91-d0e0-cd0b23923196" [ 902.732991] env[62235]: _type = "Task" [ 902.732991] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.756824] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271867, 'name': PowerOffVM_Task, 'duration_secs': 0.638239} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.757220] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5263fc0d-7bd9-2e91-d0e0-cd0b23923196, 'name': SearchDatastore_Task, 'duration_secs': 0.013725} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.757525] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.758852] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f56d926-406a-48e3-b538-6e472c2d6b89 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.761507] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61681a8e-015d-45cf-95f1-27d2e095b79d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.771367] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 902.771367] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523255f3-febb-ebde-d135-30e21242eb25" [ 902.771367] env[62235]: _type = "Task" [ 902.771367] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.794761] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc1096e-623a-455a-9cc0-40c1bf10b53c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.805195] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523255f3-febb-ebde-d135-30e21242eb25, 'name': SearchDatastore_Task, 'duration_secs': 0.012623} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.807679] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.809445] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4/3ba0a88f-6de9-46ed-9c98-306e13dcc3f4.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.814562] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-367addb3-b2ab-4386-b7bc-7412d4ab71d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.824905] env[62235]: DEBUG nova.objects.instance [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'flavor' on Instance uuid 15d41c0f-045d-4e1c-88b5-1841f99f2e74 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.826926] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 902.826926] env[62235]: value = "task-1271871" [ 902.826926] env[62235]: _type = "Task" [ 902.826926] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.839621] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.934206] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf6629b-4534-4113-afd2-8e6beb9f5468 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.945651] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e810eb-cbb3-47a0-9f2d-8e55a40bc0a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.990498] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb65d32-60eb-4e6f-a741-ed1804cd3a51 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.000659] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26410a1e-7306-4dd6-8b60-504d41d57c89 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.018890] env[62235]: DEBUG nova.compute.provider_tree [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.128836] env[62235]: DEBUG oslo_concurrency.lockutils [None req-099962bd-adec-4562-a58a-88d9607049e8 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "b5e4c6b2-4583-4a20-b570-99699063e172" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.017s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.293787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.293787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.294058] env[62235]: DEBUG nova.network.neutron [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.294165] env[62235]: DEBUG nova.objects.instance [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lazy-loading 'info_cache' on Instance uuid 44ed8739-ad20-41a3-abe1-38176c1bee25 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.313485] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 903.314843] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-258e915d-c0bd-434f-8888-3d21960960a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.327251] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 903.327251] env[62235]: value = "task-1271872" [ 903.327251] env[62235]: _type = "Task" [ 903.327251] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.337412] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0fcf659-556f-4e9a-96a3-d6850ecab511 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.341s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.359547] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271872, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.360769] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271871, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.522856] env[62235]: DEBUG nova.scheduler.client.report [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.562609] env[62235]: DEBUG nova.network.neutron [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Updated VIF entry in instance network info cache for port 62b6b792-dba7-4eaf-b0d7-715fc1e84270. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.562881] env[62235]: DEBUG nova.network.neutron [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Updating instance_info_cache with network_info: [{"id": "62b6b792-dba7-4eaf-b0d7-715fc1e84270", "address": "fa:16:3e:22:ad:a5", "network": {"id": "fc237f42-52cc-4043-8331-d796c72abb6c", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1222472131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e63e0ba8348ab917d44a623f8d117", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62b6b792-db", "ovs_interfaceid": "62b6b792-dba7-4eaf-b0d7-715fc1e84270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.716100] env[62235]: DEBUG nova.compute.manager [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.717250] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403caab7-a205-4a59-8c2c-9a5f3f1c390d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.842275] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271872, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.845867] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596296} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.846190] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4/3ba0a88f-6de9-46ed-9c98-306e13dcc3f4.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.846417] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.846700] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23f0afbf-2f93-4eaf-9b0f-3b54a135e5e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.855483] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 903.855483] env[62235]: value = "task-1271873" [ 903.855483] env[62235]: _type = "Task" [ 903.855483] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.866808] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271873, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.029858] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.029858] env[62235]: DEBUG nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.032371] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.064s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.032610] env[62235]: DEBUG nova.objects.instance [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lazy-loading 'resources' on Instance uuid 06ec6b9c-bef0-4f96-a185-4315961de7f7 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.066196] env[62235]: DEBUG oslo_concurrency.lockutils [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] Releasing lock "refresh_cache-3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.066491] env[62235]: DEBUG nova.compute.manager [req-e4a0bae6-7b48-4720-917a-6a4ddf9800af req-e66b359a-4ac7-4535-b0eb-1dc1497bac68 service nova] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Received event network-vif-deleted-0d781479-5511-42b0-9aa5-4efbe57d8d58 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.229088] env[62235]: INFO nova.compute.manager [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] instance snapshotting [ 904.232831] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fc54e0-2d04-4c3b-b804-9aa3fe7670be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.254000] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d93763-cfed-4071-8ae6-46833e1b75d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.339692] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271872, 'name': CreateSnapshot_Task, 'duration_secs': 0.609655} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.339850] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 904.343354] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5348d5-ecb4-466a-96b8-548e6b67290b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.365407] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271873, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0929} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.366284] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.367281] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c13f1bb-15f1-410c-bf2f-bcd042c6dd75 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.391992] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4/3ba0a88f-6de9-46ed-9c98-306e13dcc3f4.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.393080] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48e649e1-5268-4edf-bd6a-e2a43f6596d9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.419188] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 904.419188] env[62235]: value = "task-1271874" [ 904.419188] env[62235]: _type = "Task" [ 904.419188] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.430172] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271874, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.537175] env[62235]: DEBUG nova.compute.utils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.541658] env[62235]: DEBUG nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.541889] env[62235]: DEBUG nova.network.neutron [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.640641] env[62235]: DEBUG nova.policy [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6bf04a20a5a44037bc491831e0db83f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'baccc8328255412693198d697f82cdf7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.685963] env[62235]: DEBUG nova.network.neutron [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance_info_cache with network_info: [{"id": "3c9265cd-f975-407d-8f23-ab587b57e089", "address": "fa:16:3e:b3:9f:3c", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c9265cd-f9", "ovs_interfaceid": "3c9265cd-f975-407d-8f23-ab587b57e089", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.719568] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5186ba-528a-4e7d-b1d5-2d50516dbe20 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.728788] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ecab5f-eac4-4ae2-b505-a6d0d5b80b25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.768898] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 904.769927] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5054e786-67b5-4c70-91d6-002e4ce20db7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.772615] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21249084-ff29-4b69-9daf-bea9947dae88 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.781367] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a248105-26ca-4f45-a98e-cf9baa2cbddb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.789242] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 904.789242] env[62235]: value = "task-1271875" [ 904.789242] env[62235]: _type = "Task" [ 904.789242] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.801505] env[62235]: DEBUG nova.compute.provider_tree [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.808794] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271875, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.862790] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 904.863494] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d3381e91-fe86-4888-9400-c78b70d0777f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.875132] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 904.875132] env[62235]: value = "task-1271876" [ 904.875132] env[62235]: _type = "Task" [ 904.875132] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.883918] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271876, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.936271] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271874, 'name': ReconfigVM_Task, 'duration_secs': 0.385324} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.936594] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4/3ba0a88f-6de9-46ed-9c98-306e13dcc3f4.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.937325] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d16234b3-a6b8-401f-b072-cd41b0a5fca8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.949751] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 904.949751] env[62235]: value = "task-1271877" [ 904.949751] env[62235]: _type = "Task" [ 904.949751] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.962256] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271877, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.042601] env[62235]: DEBUG nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.086653] env[62235]: DEBUG nova.network.neutron [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Successfully created port: e66e8da7-6e73-460a-8812-12398d18bd9c {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.105342] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.105626] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.142268] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52df9530-d6af-58e4-c81a-51066ce5190b/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 905.143382] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e67ba7-4031-4e7e-b0be-09f50f7b3f77 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.150172] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52df9530-d6af-58e4-c81a-51066ce5190b/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 905.150350] env[62235]: ERROR oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52df9530-d6af-58e4-c81a-51066ce5190b/disk-0.vmdk due to incomplete transfer. [ 905.150577] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e1e80a90-69ae-4792-a083-9379cc6ec7ca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.158196] env[62235]: DEBUG oslo_vmware.rw_handles [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52df9530-d6af-58e4-c81a-51066ce5190b/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 905.158430] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Uploaded image 7e6294f3-c087-4ce3-bd45-09f6dccc5c97 to the Glance image server {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 905.160808] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 905.161107] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-139e458c-d2dd-46fe-8a0e-f6ac90f321f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.171035] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 905.171035] env[62235]: value = "task-1271878" [ 905.171035] env[62235]: _type = "Task" [ 905.171035] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.181131] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271878, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.190328] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-44ed8739-ad20-41a3-abe1-38176c1bee25" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.190633] env[62235]: DEBUG nova.objects.instance [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lazy-loading 'migration_context' on Instance uuid 44ed8739-ad20-41a3-abe1-38176c1bee25 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.301518] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271875, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.305614] env[62235]: DEBUG nova.scheduler.client.report [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.351780] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "2068403d-5870-4daa-b4d6-acaf39209488" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.352201] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "2068403d-5870-4daa-b4d6-acaf39209488" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.385563] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271876, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.460018] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271877, 'name': Rename_Task, 'duration_secs': 0.40391} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.460395] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.460675] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cb070aa-82a4-41af-aaea-876bfba264ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.467988] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 905.467988] env[62235]: value = "task-1271879" [ 905.467988] env[62235]: _type = "Task" [ 905.467988] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.476685] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271879, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.608033] env[62235]: DEBUG nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 905.682665] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271878, 'name': Destroy_Task} progress is 33%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.693885] env[62235]: DEBUG nova.objects.base [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Object Instance<44ed8739-ad20-41a3-abe1-38176c1bee25> lazy-loaded attributes: info_cache,migration_context {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 905.694820] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bb3385-45e7-4739-a92c-98225d05b014 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.716947] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5045b12a-520c-4ac1-b448-dcbc5b83a494 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.723394] env[62235]: DEBUG oslo_vmware.api [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 905.723394] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523ef324-564f-12d3-a6dd-af99bff97e13" [ 905.723394] env[62235]: _type = "Task" [ 905.723394] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.732203] env[62235]: DEBUG oslo_vmware.api [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523ef324-564f-12d3-a6dd-af99bff97e13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.801975] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271875, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.810980] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.834756] env[62235]: INFO nova.scheduler.client.report [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Deleted allocations for instance 06ec6b9c-bef0-4f96-a185-4315961de7f7 [ 905.855477] env[62235]: DEBUG nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 905.885980] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271876, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.981519] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271879, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.059738] env[62235]: DEBUG nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.080964] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.081227] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.081398] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.081625] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.081792] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.081948] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.082182] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.082354] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.082548] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.082689] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.082905] env[62235]: DEBUG nova.virt.hardware [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.083773] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777f9f0f-a34c-4da0-bef7-6c07e0379b95 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.092169] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fd71b8-14d9-493e-98f1-a919409883e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.126033] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.126136] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.127730] env[62235]: INFO nova.compute.claims [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.183332] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271878, 'name': Destroy_Task, 'duration_secs': 0.673433} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.183591] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Destroyed the VM [ 906.183853] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 906.184146] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-36ac68e7-7a2a-4173-a1f8-71b40412d3af {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.192528] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 906.192528] env[62235]: value = "task-1271880" [ 906.192528] env[62235]: _type = "Task" [ 906.192528] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.201078] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271880, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.234986] env[62235]: DEBUG oslo_vmware.api [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523ef324-564f-12d3-a6dd-af99bff97e13, 'name': SearchDatastore_Task, 'duration_secs': 0.01183} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.235753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.302479] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271875, 'name': CreateSnapshot_Task, 'duration_secs': 1.235134} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.302782] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 906.303530] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb2bd7e-844e-4e44-94c6-df6b11bfe480 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.342984] env[62235]: DEBUG oslo_concurrency.lockutils [None req-469a5651-68e7-402d-840f-db0394911b67 tempest-AttachVolumeTestJSON-330186742 tempest-AttachVolumeTestJSON-330186742-project-member] Lock "06ec6b9c-bef0-4f96-a185-4315961de7f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.807s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.377119] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.389540] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271876, 'name': CloneVM_Task} progress is 95%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.481168] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271879, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.618681] env[62235]: DEBUG nova.compute.manager [req-1bc0bb99-d9d3-4575-923a-615d3faa6736 req-4c142070-ef10-477d-91af-c6076d00e6bc service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Received event network-vif-plugged-e66e8da7-6e73-460a-8812-12398d18bd9c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.618950] env[62235]: DEBUG oslo_concurrency.lockutils [req-1bc0bb99-d9d3-4575-923a-615d3faa6736 req-4c142070-ef10-477d-91af-c6076d00e6bc service nova] Acquiring lock "ecb104db-2172-4b09-867a-7a2813e03fa5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.619206] env[62235]: DEBUG oslo_concurrency.lockutils [req-1bc0bb99-d9d3-4575-923a-615d3faa6736 req-4c142070-ef10-477d-91af-c6076d00e6bc service nova] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.619431] env[62235]: DEBUG oslo_concurrency.lockutils [req-1bc0bb99-d9d3-4575-923a-615d3faa6736 req-4c142070-ef10-477d-91af-c6076d00e6bc service nova] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.619615] env[62235]: DEBUG nova.compute.manager [req-1bc0bb99-d9d3-4575-923a-615d3faa6736 req-4c142070-ef10-477d-91af-c6076d00e6bc service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] No waiting events found dispatching network-vif-plugged-e66e8da7-6e73-460a-8812-12398d18bd9c {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.619826] env[62235]: WARNING nova.compute.manager [req-1bc0bb99-d9d3-4575-923a-615d3faa6736 req-4c142070-ef10-477d-91af-c6076d00e6bc service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Received unexpected event network-vif-plugged-e66e8da7-6e73-460a-8812-12398d18bd9c for instance with vm_state building and task_state spawning. [ 906.681895] env[62235]: DEBUG nova.network.neutron [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Successfully updated port: e66e8da7-6e73-460a-8812-12398d18bd9c {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.704728] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271880, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.822232] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 906.822571] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-28e06065-ab15-47ca-bdc0-f4497d0ac68e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.834026] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 906.834026] env[62235]: value = "task-1271882" [ 906.834026] env[62235]: _type = "Task" [ 906.834026] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.842831] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271882, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.888439] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271876, 'name': CloneVM_Task, 'duration_secs': 1.714507} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.888740] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Created linked-clone VM from snapshot [ 906.889601] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c182e84-7e0d-476b-b22a-7f8047f98fdd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.898724] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Uploading image be3e2592-5366-4a7a-b8e7-a28c16a9ddf3 {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 906.927918] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 906.927918] env[62235]: value = "vm-273478" [ 906.927918] env[62235]: _type = "VirtualMachine" [ 906.927918] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 906.928256] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0cf8d021-69fd-43d5-9b39-6c05d07a6d28 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.938163] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lease: (returnval){ [ 906.938163] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ae150-15a8-ade5-8669-6894f27dac06" [ 906.938163] env[62235]: _type = "HttpNfcLease" [ 906.938163] env[62235]: } obtained for exporting VM: (result){ [ 906.938163] env[62235]: value = "vm-273478" [ 906.938163] env[62235]: _type = "VirtualMachine" [ 906.938163] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 906.938572] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the lease: (returnval){ [ 906.938572] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ae150-15a8-ade5-8669-6894f27dac06" [ 906.938572] env[62235]: _type = "HttpNfcLease" [ 906.938572] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 906.948357] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 906.948357] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ae150-15a8-ade5-8669-6894f27dac06" [ 906.948357] env[62235]: _type = "HttpNfcLease" [ 906.948357] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 906.981208] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271879, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.183195] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.183355] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquired lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.183515] env[62235]: DEBUG nova.network.neutron [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.207682] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271880, 'name': RemoveSnapshot_Task} progress is 56%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.295791] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241abc37-a946-459e-a427-2fc40d55a9b3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.304901] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82219086-5a8f-491d-9bd5-4989f4b1171b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.339527] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e860c0e9-3239-48e4-8632-82f451dcc8e4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.348567] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271882, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.351702] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8cbb66-7db5-4c25-971b-31ef4227159d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.367849] env[62235]: DEBUG nova.compute.provider_tree [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.450132] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 907.450132] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ae150-15a8-ade5-8669-6894f27dac06" [ 907.450132] env[62235]: _type = "HttpNfcLease" [ 907.450132] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 907.450568] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 907.450568] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ae150-15a8-ade5-8669-6894f27dac06" [ 907.450568] env[62235]: _type = "HttpNfcLease" [ 907.450568] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 907.451662] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd84e2c-1f3e-4d6d-b4d2-08edd9fc80ee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.462103] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cb3fa8-c92a-2ac8-4095-6c5177b3c382/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 907.462392] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cb3fa8-c92a-2ac8-4095-6c5177b3c382/disk-0.vmdk for reading. {{(pid=62235) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 907.539705] env[62235]: DEBUG oslo_vmware.api [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271879, 'name': PowerOnVM_Task, 'duration_secs': 1.57347} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.539978] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.540211] env[62235]: INFO nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Took 7.98 seconds to spawn the instance on the hypervisor. [ 907.540414] env[62235]: DEBUG nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.542096] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b017176a-760b-49f6-91a6-af9e4bbda6cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.576027] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0ccdf6f3-4b56-4270-ba66-ba01774290df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.704568] env[62235]: DEBUG oslo_vmware.api [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271880, 'name': RemoveSnapshot_Task, 'duration_secs': 1.386264} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.704859] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 907.705421] env[62235]: INFO nova.compute.manager [None req-d018ff61-5f7a-44a9-a2fe-0c4a4d0db6d9 tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Took 16.51 seconds to snapshot the instance on the hypervisor. [ 907.749424] env[62235]: DEBUG nova.network.neutron [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.847624] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271882, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.870739] env[62235]: DEBUG nova.scheduler.client.report [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.936446] env[62235]: DEBUG nova.network.neutron [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Updating instance_info_cache with network_info: [{"id": "e66e8da7-6e73-460a-8812-12398d18bd9c", "address": "fa:16:3e:d6:61:e0", "network": {"id": "095f993a-0f79-4100-ab40-90d159a2bacf", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1486577542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "baccc8328255412693198d697f82cdf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ef6889-a40c-40f5-a6e5-d8726606296a", "external-id": "nsx-vlan-transportzone-537", "segmentation_id": 537, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape66e8da7-6e", "ovs_interfaceid": "e66e8da7-6e73-460a-8812-12398d18bd9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.061750] env[62235]: INFO nova.compute.manager [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Took 27.24 seconds to build instance. [ 908.348946] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271882, 'name': CloneVM_Task, 'duration_secs': 1.373549} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.348946] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Created linked-clone VM from snapshot [ 908.350124] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a2d300-f4bb-4fd9-a644-4728bd15f0cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.358753] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Uploading image 35cb92c4-d716-4498-9ddc-853518ba56f2 {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 908.370387] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 908.370655] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5aaa259a-3022-4a6d-842d-f9655e80fb99 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.375818] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.376447] env[62235]: DEBUG nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.383542] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.148s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.385022] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 908.385022] env[62235]: value = "task-1271884" [ 908.385022] env[62235]: _type = "Task" [ 908.385022] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.396531] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271884, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.439964] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Releasing lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.440345] env[62235]: DEBUG nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Instance network_info: |[{"id": "e66e8da7-6e73-460a-8812-12398d18bd9c", "address": "fa:16:3e:d6:61:e0", "network": {"id": "095f993a-0f79-4100-ab40-90d159a2bacf", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1486577542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "baccc8328255412693198d697f82cdf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ef6889-a40c-40f5-a6e5-d8726606296a", "external-id": "nsx-vlan-transportzone-537", "segmentation_id": 537, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape66e8da7-6e", "ovs_interfaceid": "e66e8da7-6e73-460a-8812-12398d18bd9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.440891] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:61:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53ef6889-a40c-40f5-a6e5-d8726606296a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e66e8da7-6e73-460a-8812-12398d18bd9c', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.449541] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Creating folder: Project (baccc8328255412693198d697f82cdf7). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.450595] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dee0c1a4-674c-4b45-b70f-b6ad458905ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.464612] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Created folder: Project (baccc8328255412693198d697f82cdf7) in parent group-v273362. [ 908.464612] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Creating folder: Instances. Parent ref: group-v273481. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.464612] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04594c46-1d62-42cd-a567-3f774b9d58a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.479652] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Created folder: Instances in parent group-v273481. [ 908.480212] env[62235]: DEBUG oslo.service.loopingcall [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.480516] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.480775] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19551df7-46e9-4c57-aac0-6c04bacd917b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.503351] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.503351] env[62235]: value = "task-1271887" [ 908.503351] env[62235]: _type = "Task" [ 908.503351] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.513289] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271887, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.564031] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ea8811b7-6966-4c2e-91dd-4259e63edb37 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.746s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.662275] env[62235]: DEBUG nova.compute.manager [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Received event network-changed-e66e8da7-6e73-460a-8812-12398d18bd9c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.662794] env[62235]: DEBUG nova.compute.manager [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Refreshing instance network info cache due to event network-changed-e66e8da7-6e73-460a-8812-12398d18bd9c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.662794] env[62235]: DEBUG oslo_concurrency.lockutils [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] Acquiring lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.663094] env[62235]: DEBUG oslo_concurrency.lockutils [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] Acquired lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.663463] env[62235]: DEBUG nova.network.neutron [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Refreshing network info cache for port e66e8da7-6e73-460a-8812-12398d18bd9c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.889136] env[62235]: DEBUG nova.compute.utils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.889880] env[62235]: DEBUG nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.890202] env[62235]: DEBUG nova.network.neutron [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.911805] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271884, 'name': Destroy_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.912466] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.915030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.915030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.915030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.915030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.921552] env[62235]: INFO nova.compute.manager [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Terminating instance [ 908.925867] env[62235]: DEBUG nova.compute.manager [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.926415] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 908.927430] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f16361-1655-4743-b664-7c26dae50e4f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.943634] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.944859] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b98db62-d090-49a5-9736-a5f6a58aeafa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.957605] env[62235]: DEBUG oslo_vmware.api [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 908.957605] env[62235]: value = "task-1271888" [ 908.957605] env[62235]: _type = "Task" [ 908.957605] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.972365] env[62235]: DEBUG oslo_vmware.api [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271888, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.990740] env[62235]: DEBUG nova.policy [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8a1c9e076304498999fe4c6c1d89c69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a272872fcc55419fb474121b7fefd4c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.018976] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271887, 'name': CreateVM_Task, 'duration_secs': 0.443062} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.021395] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 909.022298] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.022524] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.023076] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 909.023261] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16b10dc7-37e3-427a-923e-ed3644f702b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.030650] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 909.030650] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52575198-2f06-5d6b-6f62-d809a96f9a1e" [ 909.030650] env[62235]: _type = "Task" [ 909.030650] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.042116] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52575198-2f06-5d6b-6f62-d809a96f9a1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.141219] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f21322-79c3-422f-8efe-1ef1ba28a4f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.150526] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7e938f-a219-4f1a-99b2-6453c91477f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.185989] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fddb3fe-dfad-4281-8563-b267c7c23647 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.195544] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9c0388-b2a3-4330-a552-8aba2e20afea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.210676] env[62235]: DEBUG nova.compute.provider_tree [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.286383] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "507fa559-1889-4f6e-81ee-93eb14d134b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.286651] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.286867] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "507fa559-1889-4f6e-81ee-93eb14d134b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.287063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.287247] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.289688] env[62235]: INFO nova.compute.manager [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Terminating instance [ 909.291782] env[62235]: DEBUG nova.compute.manager [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.292053] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.292841] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f000b84e-741c-463b-be27-ebd8deda27d5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.301455] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.301733] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-953834af-c9f7-4845-8971-efde1e5425ed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.312063] env[62235]: DEBUG oslo_vmware.api [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 909.312063] env[62235]: value = "task-1271889" [ 909.312063] env[62235]: _type = "Task" [ 909.312063] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.321908] env[62235]: DEBUG oslo_vmware.api [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.399924] env[62235]: DEBUG nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.404138] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271884, 'name': Destroy_Task, 'duration_secs': 0.529301} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.404552] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Destroyed the VM [ 909.405108] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 909.405354] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a8d30bdc-3562-42a1-b551-9cc742f228b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.415563] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 909.415563] env[62235]: value = "task-1271890" [ 909.415563] env[62235]: _type = "Task" [ 909.415563] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.427505] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271890, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.474457] env[62235]: DEBUG oslo_vmware.api [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271888, 'name': PowerOffVM_Task, 'duration_secs': 0.217873} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.478777] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.479140] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.479496] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdd8bc3f-cafa-4de4-8cfa-864b8991d23a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.548758] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52575198-2f06-5d6b-6f62-d809a96f9a1e, 'name': SearchDatastore_Task, 'duration_secs': 0.016134} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.548880] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.550092] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.550092] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.550092] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.550092] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.551191] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36f190f1-f32f-420a-82d9-05545fb33828 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.553253] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.553464] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.553664] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Deleting the datastore file [datastore1] 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.553938] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-179efdb1-feb9-4012-ba20-1efc8c39f638 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.562528] env[62235]: DEBUG oslo_vmware.api [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for the task: (returnval){ [ 909.562528] env[62235]: value = "task-1271892" [ 909.562528] env[62235]: _type = "Task" [ 909.562528] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.567523] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.567763] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.568954] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcaea9de-4484-43ff-b43f-a675b0776f2d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.574488] env[62235]: DEBUG oslo_vmware.api [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271892, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.578673] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 909.578673] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cbc63c-632f-7446-4aaa-7dd1e6798eb2" [ 909.578673] env[62235]: _type = "Task" [ 909.578673] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.587868] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cbc63c-632f-7446-4aaa-7dd1e6798eb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.623438] env[62235]: DEBUG nova.network.neutron [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Successfully created port: 984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.659758] env[62235]: DEBUG nova.network.neutron [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Updated VIF entry in instance network info cache for port e66e8da7-6e73-460a-8812-12398d18bd9c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.660216] env[62235]: DEBUG nova.network.neutron [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Updating instance_info_cache with network_info: [{"id": "e66e8da7-6e73-460a-8812-12398d18bd9c", "address": "fa:16:3e:d6:61:e0", "network": {"id": "095f993a-0f79-4100-ab40-90d159a2bacf", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1486577542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "baccc8328255412693198d697f82cdf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ef6889-a40c-40f5-a6e5-d8726606296a", "external-id": "nsx-vlan-transportzone-537", "segmentation_id": 537, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape66e8da7-6e", "ovs_interfaceid": "e66e8da7-6e73-460a-8812-12398d18bd9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.716343] env[62235]: DEBUG nova.scheduler.client.report [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.828039] env[62235]: DEBUG oslo_vmware.api [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271889, 'name': PowerOffVM_Task, 'duration_secs': 0.198313} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.828039] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.828039] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.828039] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34d3190e-97cb-4b67-b11b-39204606d281 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.921023] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.921023] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.921023] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Deleting the datastore file [datastore2] 507fa559-1889-4f6e-81ee-93eb14d134b2 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.924720] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49e58bd9-bbc0-4138-9d4f-1d1e37b182dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.934503] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271890, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.937023] env[62235]: DEBUG oslo_vmware.api [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for the task: (returnval){ [ 909.937023] env[62235]: value = "task-1271895" [ 909.937023] env[62235]: _type = "Task" [ 909.937023] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.948562] env[62235]: DEBUG oslo_vmware.api [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.073685] env[62235]: DEBUG oslo_vmware.api [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Task: {'id': task-1271892, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217669} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.074007] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.074224] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.074409] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.074618] env[62235]: INFO nova.compute.manager [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 910.074931] env[62235]: DEBUG oslo.service.loopingcall [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.075182] env[62235]: DEBUG nova.compute.manager [-] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.075288] env[62235]: DEBUG nova.network.neutron [-] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 910.089583] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cbc63c-632f-7446-4aaa-7dd1e6798eb2, 'name': SearchDatastore_Task, 'duration_secs': 0.02072} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.090507] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-178374ed-5c28-4eef-9e0b-c013bcc50eaa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.096554] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 910.096554] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ef4d0e-fc64-b711-ba88-dd641d3a7f4e" [ 910.096554] env[62235]: _type = "Task" [ 910.096554] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.105353] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ef4d0e-fc64-b711-ba88-dd641d3a7f4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.163649] env[62235]: DEBUG oslo_concurrency.lockutils [req-777bc19a-bcb3-419c-b82b-3602c7c4376b req-77ec0a5f-6fe7-40de-a53a-7ba8118d9602 service nova] Releasing lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.413166] env[62235]: DEBUG nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.432090] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271890, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.446625] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.446935] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.447119] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.447333] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.447476] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.447631] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.447845] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.448190] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.448404] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.448581] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.448764] env[62235]: DEBUG nova.virt.hardware [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.449654] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37643def-64ae-4ea5-a008-2fe9e7b507d0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.458655] env[62235]: DEBUG oslo_vmware.api [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Task: {'id': task-1271895, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259703} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.459367] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.459598] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.459806] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.459987] env[62235]: INFO nova.compute.manager [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Took 1.17 seconds to destroy the instance on the hypervisor. [ 910.460250] env[62235]: DEBUG oslo.service.loopingcall [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.462261] env[62235]: DEBUG nova.compute.manager [-] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.462371] env[62235]: DEBUG nova.network.neutron [-] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 910.467480] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6ea699-303f-4e2d-862e-50594aca45da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.608987] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ef4d0e-fc64-b711-ba88-dd641d3a7f4e, 'name': SearchDatastore_Task, 'duration_secs': 0.026151} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.609414] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.609703] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] ecb104db-2172-4b09-867a-7a2813e03fa5/ecb104db-2172-4b09-867a-7a2813e03fa5.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.609974] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-081fd4e2-cd0d-4f95-8595-e90c11277b6a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.617858] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 910.617858] env[62235]: value = "task-1271896" [ 910.617858] env[62235]: _type = "Task" [ 910.617858] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.627643] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.689027] env[62235]: DEBUG nova.compute.manager [req-1848accd-3cdb-4891-9252-72d8d4677fa0 req-5958457d-6678-40d0-96c0-2116d510d838 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Received event network-vif-deleted-62b6b792-dba7-4eaf-b0d7-715fc1e84270 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.689263] env[62235]: INFO nova.compute.manager [req-1848accd-3cdb-4891-9252-72d8d4677fa0 req-5958457d-6678-40d0-96c0-2116d510d838 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Neutron deleted interface 62b6b792-dba7-4eaf-b0d7-715fc1e84270; detaching it from the instance and deleting it from the info cache [ 910.689475] env[62235]: DEBUG nova.network.neutron [req-1848accd-3cdb-4891-9252-72d8d4677fa0 req-5958457d-6678-40d0-96c0-2116d510d838 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.729114] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.345s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.732864] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.356s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.735357] env[62235]: INFO nova.compute.claims [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.906042] env[62235]: DEBUG nova.network.neutron [-] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.932394] env[62235]: DEBUG oslo_vmware.api [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271890, 'name': RemoveSnapshot_Task, 'duration_secs': 1.119512} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.932732] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 911.129967] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271896, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.184144] env[62235]: DEBUG nova.network.neutron [-] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.199024] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-789b7b12-b75a-4b35-956e-4a5e17061e13 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.212020] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cfa0ac-e13d-4169-b7a3-9c47c538baef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.228963] env[62235]: DEBUG nova.compute.manager [req-6fb7f1f8-f488-4bc3-b7f6-cf3f4d62aee3 req-253e781a-18a3-40d1-98df-811785d12de2 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-vif-plugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.228963] env[62235]: DEBUG oslo_concurrency.lockutils [req-6fb7f1f8-f488-4bc3-b7f6-cf3f4d62aee3 req-253e781a-18a3-40d1-98df-811785d12de2 service nova] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.228963] env[62235]: DEBUG oslo_concurrency.lockutils [req-6fb7f1f8-f488-4bc3-b7f6-cf3f4d62aee3 req-253e781a-18a3-40d1-98df-811785d12de2 service nova] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.228963] env[62235]: DEBUG oslo_concurrency.lockutils [req-6fb7f1f8-f488-4bc3-b7f6-cf3f4d62aee3 req-253e781a-18a3-40d1-98df-811785d12de2 service nova] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.228963] env[62235]: DEBUG nova.compute.manager [req-6fb7f1f8-f488-4bc3-b7f6-cf3f4d62aee3 req-253e781a-18a3-40d1-98df-811785d12de2 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] No waiting events found dispatching network-vif-plugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.228963] env[62235]: WARNING nova.compute.manager [req-6fb7f1f8-f488-4bc3-b7f6-cf3f4d62aee3 req-253e781a-18a3-40d1-98df-811785d12de2 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received unexpected event network-vif-plugged-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 for instance with vm_state shelved_offloaded and task_state spawning. [ 911.255294] env[62235]: DEBUG nova.compute.manager [req-1848accd-3cdb-4891-9252-72d8d4677fa0 req-5958457d-6678-40d0-96c0-2116d510d838 service nova] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Detach interface failed, port_id=62b6b792-dba7-4eaf-b0d7-715fc1e84270, reason: Instance 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 911.320232] env[62235]: INFO nova.scheduler.client.report [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted allocation for migration 18afe82b-d5ea-43a2-9aad-5add4f0c891b [ 911.330545] env[62235]: DEBUG nova.network.neutron [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Successfully updated port: 984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.358134] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.358244] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.358430] env[62235]: DEBUG nova.network.neutron [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.408620] env[62235]: INFO nova.compute.manager [-] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Took 1.33 seconds to deallocate network for instance. [ 911.441102] env[62235]: WARNING nova.compute.manager [None req-c52006aa-8673-464d-832a-fefc77253702 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Image not found during snapshot: nova.exception.ImageNotFound: Image 35cb92c4-d716-4498-9ddc-853518ba56f2 could not be found. [ 911.629710] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271896, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652709} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.630048] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] ecb104db-2172-4b09-867a-7a2813e03fa5/ecb104db-2172-4b09-867a-7a2813e03fa5.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.630260] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.630573] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-442a79ed-d495-46da-9c4e-7722d9f2de30 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.640922] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 911.640922] env[62235]: value = "task-1271897" [ 911.640922] env[62235]: _type = "Task" [ 911.640922] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.652281] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.685711] env[62235]: INFO nova.compute.manager [-] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Took 1.22 seconds to deallocate network for instance. [ 911.828988] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6e0b8a9f-45ed-487e-8af7-76a23708a3f6 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.139s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.833525] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.833525] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.833525] env[62235]: DEBUG nova.network.neutron [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.917816] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa718df-37f5-48e6-a910-310a382ef2b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.921435] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.927650] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc1221d-43e2-4fc2-be12-39cb004511a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.967780] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f65fcc5-5a7e-4145-a439-6b7341166fce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.977482] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2ef2f7-12d0-434a-b386-3f7e9affb91f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.993174] env[62235]: DEBUG nova.compute.provider_tree [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.009659] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "b5e4c6b2-4583-4a20-b570-99699063e172" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.009995] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "b5e4c6b2-4583-4a20-b570-99699063e172" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.010304] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "b5e4c6b2-4583-4a20-b570-99699063e172-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.010524] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "b5e4c6b2-4583-4a20-b570-99699063e172-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.010716] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "b5e4c6b2-4583-4a20-b570-99699063e172-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.013190] env[62235]: INFO nova.compute.manager [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Terminating instance [ 912.015596] env[62235]: DEBUG nova.compute.manager [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 912.015738] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.016595] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbe2094-db02-42f1-b309-03adf2663fb8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.027887] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.028264] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fe7a9d6-0a93-411f-a4c4-8d207bfd34df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.041139] env[62235]: DEBUG oslo_vmware.api [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 912.041139] env[62235]: value = "task-1271898" [ 912.041139] env[62235]: _type = "Task" [ 912.041139] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.055886] env[62235]: DEBUG oslo_vmware.api [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271898, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.154244] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069246} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.154483] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 912.155363] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa160d2-a604-4f26-9fbe-fa5d92398091 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.179739] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] ecb104db-2172-4b09-867a-7a2813e03fa5/ecb104db-2172-4b09-867a-7a2813e03fa5.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.180091] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d901e349-65f9-4563-9866-12c48fe1ed7e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.195118] env[62235]: DEBUG nova.network.neutron [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.197641] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.205042] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 912.205042] env[62235]: value = "task-1271899" [ 912.205042] env[62235]: _type = "Task" [ 912.205042] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.213940] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271899, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.375624] env[62235]: DEBUG nova.network.neutron [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.496828] env[62235]: DEBUG nova.scheduler.client.report [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.517413] env[62235]: DEBUG nova.network.neutron [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Updating instance_info_cache with network_info: [{"id": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "address": "fa:16:3e:fa:ee:90", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984bbc0d-59", "ovs_interfaceid": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.552727] env[62235]: DEBUG oslo_vmware.api [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271898, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.698770] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.716457] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271899, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.727719] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='affc41485e69dd349e2a48e93b8d7045',container_format='bare',created_at=2024-10-12T06:24:00Z,direct_url=,disk_format='vmdk',id=0b74899a-987d-4936-af42-5d58c7e24984,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1539193385-shelved',owner='6832a93601f84138aeed5a906f24394b',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-12T06:24:15Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.728067] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.728324] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.728904] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.728904] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.728904] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.729118] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.729308] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.729510] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.729700] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.729881] env[62235]: DEBUG nova.virt.hardware [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.730829] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a09c9a-4991-462a-8e7c-f3d11b777b65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.740144] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de41be48-e36a-4493-8400-24c7d9c3f947 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.755117] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:60:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.763388] env[62235]: DEBUG oslo.service.loopingcall [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.763756] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.764018] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97643e61-4380-4daa-b7e8-0c7ac9877be0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.785670] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.785670] env[62235]: value = "task-1271900" [ 912.785670] env[62235]: _type = "Task" [ 912.785670] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.794798] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271900, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.825996] env[62235]: DEBUG nova.compute.manager [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Received event network-vif-deleted-6b26570e-71cc-4c85-b5d6-d16ced84d829 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.826542] env[62235]: DEBUG nova.compute.manager [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Received event network-vif-plugged-984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.826830] env[62235]: DEBUG oslo_concurrency.lockutils [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] Acquiring lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.827130] env[62235]: DEBUG oslo_concurrency.lockutils [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.827360] env[62235]: DEBUG oslo_concurrency.lockutils [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.827572] env[62235]: DEBUG nova.compute.manager [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] No waiting events found dispatching network-vif-plugged-984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.827795] env[62235]: WARNING nova.compute.manager [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Received unexpected event network-vif-plugged-984bbc0d-59a6-46fc-bb73-bcc8454e14e3 for instance with vm_state building and task_state spawning. [ 912.828068] env[62235]: DEBUG nova.compute.manager [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Received event network-changed-984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.828278] env[62235]: DEBUG nova.compute.manager [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Refreshing instance network info cache due to event network-changed-984bbc0d-59a6-46fc-bb73-bcc8454e14e3. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 912.828483] env[62235]: DEBUG oslo_concurrency.lockutils [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] Acquiring lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.002627] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.003209] env[62235]: DEBUG nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.006009] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.085s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.007041] env[62235]: DEBUG nova.objects.instance [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lazy-loading 'resources' on Instance uuid 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.020074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.020671] env[62235]: DEBUG nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Instance network_info: |[{"id": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "address": "fa:16:3e:fa:ee:90", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984bbc0d-59", "ovs_interfaceid": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.020796] env[62235]: DEBUG oslo_concurrency.lockutils [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] Acquired lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.020950] env[62235]: DEBUG nova.network.neutron [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Refreshing network info cache for port 984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.022175] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:ee:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '984bbc0d-59a6-46fc-bb73-bcc8454e14e3', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.030742] env[62235]: DEBUG oslo.service.loopingcall [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.032059] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.032432] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1952ab6-6966-426c-ab0f-1e97397eb331 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.063960] env[62235]: DEBUG oslo_vmware.api [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271898, 'name': PowerOffVM_Task, 'duration_secs': 1.00278} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.065730] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.066040] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.066689] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.066689] env[62235]: value = "task-1271901" [ 913.066689] env[62235]: _type = "Task" [ 913.066689] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.066890] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-648450cc-ce9d-4948-9053-14115d98d5dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.077786] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271901, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.219202] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271899, 'name': ReconfigVM_Task, 'duration_secs': 0.913453} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.219202] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Reconfigured VM instance instance-00000051 to attach disk [datastore2] ecb104db-2172-4b09-867a-7a2813e03fa5/ecb104db-2172-4b09-867a-7a2813e03fa5.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.219202] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d85715d7-346e-4709-a88c-e297d4832529 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.230261] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 913.230261] env[62235]: value = "task-1271903" [ 913.230261] env[62235]: _type = "Task" [ 913.230261] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.239322] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271903, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.297754] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271900, 'name': CreateVM_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.303257] env[62235]: DEBUG nova.compute.manager [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.303257] env[62235]: DEBUG nova.compute.manager [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing instance network info cache due to event network-changed-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.303257] env[62235]: DEBUG oslo_concurrency.lockutils [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.303257] env[62235]: DEBUG oslo_concurrency.lockutils [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.303257] env[62235]: DEBUG nova.network.neutron [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Refreshing network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.399018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "44ed8739-ad20-41a3-abe1-38176c1bee25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.399018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.399018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.399018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.399018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.399976] env[62235]: INFO nova.compute.manager [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Terminating instance [ 913.401996] env[62235]: DEBUG nova.compute.manager [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.402361] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.403324] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61ab9f2-5c5a-4fd7-b69d-95168c786ab7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.412201] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.412816] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b1fa212-fdb1-4c11-a8c5-2b89dc561f66 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.421468] env[62235]: DEBUG oslo_vmware.api [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 913.421468] env[62235]: value = "task-1271904" [ 913.421468] env[62235]: _type = "Task" [ 913.421468] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.437521] env[62235]: DEBUG oslo_vmware.api [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271904, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.512084] env[62235]: DEBUG nova.compute.utils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.516022] env[62235]: DEBUG nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.516022] env[62235]: DEBUG nova.network.neutron [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.563711] env[62235]: DEBUG nova.policy [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.583438] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271901, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.721995] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812d2b98-4e7c-46d9-8e28-0cf90786c6f7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.737828] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8123e30-083f-4ad7-983f-0067c6382a6d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.747311] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271903, 'name': Rename_Task, 'duration_secs': 0.204141} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.778280] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.778915] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-efb0bd5b-e069-456c-a386-1bfcc407259e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.782202] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd1168e-4464-481f-b1b9-27914c3baa9a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.801215] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2ec2b4-29fc-40d9-a306-59af6dd29eb4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.807988] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 913.807988] env[62235]: value = "task-1271905" [ 913.807988] env[62235]: _type = "Task" [ 913.807988] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.820663] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271900, 'name': CreateVM_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.842470] env[62235]: DEBUG nova.compute.provider_tree [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.850042] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271905, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.915787] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.916051] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.936839] env[62235]: DEBUG oslo_vmware.api [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271904, 'name': PowerOffVM_Task, 'duration_secs': 0.202805} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.937282] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.937384] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.937972] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdc40738-1840-4d9d-826d-430ed2006cd2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.016044] env[62235]: DEBUG nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.028697] env[62235]: DEBUG nova.network.neutron [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Successfully created port: 64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.080348] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271901, 'name': CreateVM_Task, 'duration_secs': 0.58688} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.083262] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.084201] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.087017] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.087017] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.087017] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe3f46ca-d019-448a-a33b-71098a2674c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.091203] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 914.091203] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52791c54-d110-21c3-9fc9-86946b06d858" [ 914.091203] env[62235]: _type = "Task" [ 914.091203] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.100882] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52791c54-d110-21c3-9fc9-86946b06d858, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.102604] env[62235]: DEBUG nova.network.neutron [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Updated VIF entry in instance network info cache for port 984bbc0d-59a6-46fc-bb73-bcc8454e14e3. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 914.103088] env[62235]: DEBUG nova.network.neutron [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Updating instance_info_cache with network_info: [{"id": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "address": "fa:16:3e:fa:ee:90", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984bbc0d-59", "ovs_interfaceid": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.138918] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.139497] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.139913] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleting the datastore file [datastore1] b5e4c6b2-4583-4a20-b570-99699063e172 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.140321] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-406e6e90-2026-4e9e-9641-1259c14b530b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.155790] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.155790] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.155790] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleting the datastore file [datastore1] 44ed8739-ad20-41a3-abe1-38176c1bee25 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.157579] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db82fb80-f1dc-447a-b404-2c729fe1bd0d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.160669] env[62235]: DEBUG oslo_vmware.api [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for the task: (returnval){ [ 914.160669] env[62235]: value = "task-1271907" [ 914.160669] env[62235]: _type = "Task" [ 914.160669] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.168384] env[62235]: DEBUG oslo_vmware.api [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 914.168384] env[62235]: value = "task-1271908" [ 914.168384] env[62235]: _type = "Task" [ 914.168384] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.173814] env[62235]: DEBUG oslo_vmware.api [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.184638] env[62235]: DEBUG oslo_vmware.api [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271908, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.304030] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271900, 'name': CreateVM_Task, 'duration_secs': 1.451642} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.304030] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.304030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.304030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.304030] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.304030] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2fc9cdf-9f11-423a-a9d0-b086535a0cce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.309509] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 914.309509] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525365e6-1d46-ac2d-baf7-326174b76930" [ 914.309509] env[62235]: _type = "Task" [ 914.309509] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.319926] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525365e6-1d46-ac2d-baf7-326174b76930, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.331811] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271905, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.349033] env[62235]: DEBUG nova.scheduler.client.report [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.421024] env[62235]: DEBUG nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.423175] env[62235]: DEBUG nova.network.neutron [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updated VIF entry in instance network info cache for port f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 914.424335] env[62235]: DEBUG nova.network.neutron [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.602910] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52791c54-d110-21c3-9fc9-86946b06d858, 'name': SearchDatastore_Task, 'duration_secs': 0.035748} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.603837] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.604649] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.606162] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.606162] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.606162] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.606447] env[62235]: DEBUG oslo_concurrency.lockutils [req-b0cb122a-f720-4dce-b821-436ea787c3ef req-45428a75-f58b-4e6e-a86b-7156dc20bc5b service nova] Releasing lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.608028] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bebe09e4-c0a1-4a7e-9258-92956202da3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.619464] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.619464] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.619464] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-514c89fa-626b-4677-b52f-16655ee57c16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.626493] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 914.626493] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dcdd56-2e20-e7e2-efdd-8d14d7e1a5b5" [ 914.626493] env[62235]: _type = "Task" [ 914.626493] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.635658] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dcdd56-2e20-e7e2-efdd-8d14d7e1a5b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.671623] env[62235]: DEBUG oslo_vmware.api [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Task: {'id': task-1271907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.255977} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.672178] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.672178] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.673030] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.673030] env[62235]: INFO nova.compute.manager [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Took 2.66 seconds to destroy the instance on the hypervisor. [ 914.673030] env[62235]: DEBUG oslo.service.loopingcall [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.673030] env[62235]: DEBUG nova.compute.manager [-] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.673030] env[62235]: DEBUG nova.network.neutron [-] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 914.684862] env[62235]: DEBUG oslo_vmware.api [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271908, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25677} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.685258] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.685522] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.685781] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.686035] env[62235]: INFO nova.compute.manager [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Took 1.28 seconds to destroy the instance on the hypervisor. [ 914.686340] env[62235]: DEBUG oslo.service.loopingcall [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.686605] env[62235]: DEBUG nova.compute.manager [-] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.686996] env[62235]: DEBUG nova.network.neutron [-] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 914.821633] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.821974] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Processing image 0b74899a-987d-4936-af42-5d58c7e24984 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.822304] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984/0b74899a-987d-4936-af42-5d58c7e24984.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.822460] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984/0b74899a-987d-4936-af42-5d58c7e24984.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.822675] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.823046] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b81fe8c8-3c98-4fa3-bc6e-23f5e946877c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.834785] env[62235]: DEBUG oslo_vmware.api [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271905, 'name': PowerOnVM_Task, 'duration_secs': 0.647125} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.835147] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.835402] env[62235]: INFO nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Took 8.78 seconds to spawn the instance on the hypervisor. [ 914.835665] env[62235]: DEBUG nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.836587] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa702693-c13a-44f2-9697-ee088ab207cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.840426] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.840635] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.841896] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d2f2f37-2a06-44f7-b32d-4dacefb3306f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.853059] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.847s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.855424] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 914.855424] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d1b9cb-7b05-eaeb-ac3e-c19f7688301d" [ 914.855424] env[62235]: _type = "Task" [ 914.855424] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.856167] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.659s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.856265] env[62235]: DEBUG nova.objects.instance [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lazy-loading 'resources' on Instance uuid 507fa559-1889-4f6e-81ee-93eb14d134b2 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.867878] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d1b9cb-7b05-eaeb-ac3e-c19f7688301d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.885575] env[62235]: INFO nova.scheduler.client.report [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Deleted allocations for instance 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4 [ 914.931042] env[62235]: DEBUG oslo_concurrency.lockutils [req-61031278-50ec-4dfc-aa9b-9d9a0ca1ee1e req-11443eb4-eef6-4043-94ef-f847719dcc99 service nova] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.953675] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.026811] env[62235]: DEBUG nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 915.059784] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.060097] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.060262] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.060708] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.060708] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.060819] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.061199] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.061199] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.061304] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.061489] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.061810] env[62235]: DEBUG nova.virt.hardware [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.062608] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c953594-9edf-43ff-8e5f-bcaaaf4ad923 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.072020] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afa8ee1-c61e-410b-87a3-9b566415cbe7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.137370] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dcdd56-2e20-e7e2-efdd-8d14d7e1a5b5, 'name': SearchDatastore_Task, 'duration_secs': 0.019304} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.138277] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b70073b-efa4-4fef-be44-4579b01cf6bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.145377] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 915.145377] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b4e9eb-1908-59f1-d25e-cf5a1f7fd10c" [ 915.145377] env[62235]: _type = "Task" [ 915.145377] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.157446] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b4e9eb-1908-59f1-d25e-cf5a1f7fd10c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.238416] env[62235]: DEBUG nova.compute.manager [req-affb12b1-42e3-4d64-bc99-5dd22d14630a req-b9dff08f-cf52-43ab-a278-aeb04ae18114 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Received event network-vif-deleted-347663dc-4c1e-4e45-ab26-82dc53ace862 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.238635] env[62235]: INFO nova.compute.manager [req-affb12b1-42e3-4d64-bc99-5dd22d14630a req-b9dff08f-cf52-43ab-a278-aeb04ae18114 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Neutron deleted interface 347663dc-4c1e-4e45-ab26-82dc53ace862; detaching it from the instance and deleting it from the info cache [ 915.238832] env[62235]: DEBUG nova.network.neutron [req-affb12b1-42e3-4d64-bc99-5dd22d14630a req-b9dff08f-cf52-43ab-a278-aeb04ae18114 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.356674] env[62235]: INFO nova.compute.manager [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Took 20.66 seconds to build instance. [ 915.367993] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Preparing fetch location {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 915.368333] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Fetch image to [datastore1] OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc/OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc.vmdk {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 915.368498] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Downloading stream optimized image 0b74899a-987d-4936-af42-5d58c7e24984 to [datastore1] OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc/OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc.vmdk on the data store datastore1 as vApp {{(pid=62235) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 915.368721] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Downloading image file data 0b74899a-987d-4936-af42-5d58c7e24984 to the ESX as VM named 'OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc' {{(pid=62235) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 915.397515] env[62235]: DEBUG oslo_concurrency.lockutils [None req-96240d9c-60f9-45b9-ad45-9e2a006abac1 tempest-ServerGroupTestJSON-934028358 tempest-ServerGroupTestJSON-934028358-project-member] Lock "3ba0a88f-6de9-46ed-9c98-306e13dcc3f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.484s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.478609] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 915.478609] env[62235]: value = "resgroup-9" [ 915.478609] env[62235]: _type = "ResourcePool" [ 915.478609] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 915.479088] env[62235]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a24e48ef-ef5f-4e7a-8007-dd18f90af7b0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.504998] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease: (returnval){ [ 915.504998] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528cdff1-b214-7021-9049-33bb4a6c50e7" [ 915.504998] env[62235]: _type = "HttpNfcLease" [ 915.504998] env[62235]: } obtained for vApp import into resource pool (val){ [ 915.504998] env[62235]: value = "resgroup-9" [ 915.504998] env[62235]: _type = "ResourcePool" [ 915.504998] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 915.505319] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the lease: (returnval){ [ 915.505319] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528cdff1-b214-7021-9049-33bb4a6c50e7" [ 915.505319] env[62235]: _type = "HttpNfcLease" [ 915.505319] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 915.515435] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.515435] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528cdff1-b214-7021-9049-33bb4a6c50e7" [ 915.515435] env[62235]: _type = "HttpNfcLease" [ 915.515435] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 915.588250] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d6d2ff-b80b-4f77-8761-d87304c4ed08 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.600026] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d9a07c-c811-4907-bb09-d297bee8e483 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.644433] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf868f1-d182-4255-a843-b61224836898 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.649574] env[62235]: DEBUG nova.compute.manager [req-1bd3e418-19ca-4cd4-b171-ec4f33ca22a4 req-439e5b54-130e-4343-aee5-2a2ccb98d268 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-vif-plugged-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.649901] env[62235]: DEBUG oslo_concurrency.lockutils [req-1bd3e418-19ca-4cd4-b171-ec4f33ca22a4 req-439e5b54-130e-4343-aee5-2a2ccb98d268 service nova] Acquiring lock "2068403d-5870-4daa-b4d6-acaf39209488-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.650110] env[62235]: DEBUG oslo_concurrency.lockutils [req-1bd3e418-19ca-4cd4-b171-ec4f33ca22a4 req-439e5b54-130e-4343-aee5-2a2ccb98d268 service nova] Lock "2068403d-5870-4daa-b4d6-acaf39209488-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.650414] env[62235]: DEBUG oslo_concurrency.lockutils [req-1bd3e418-19ca-4cd4-b171-ec4f33ca22a4 req-439e5b54-130e-4343-aee5-2a2ccb98d268 service nova] Lock "2068403d-5870-4daa-b4d6-acaf39209488-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.650508] env[62235]: DEBUG nova.compute.manager [req-1bd3e418-19ca-4cd4-b171-ec4f33ca22a4 req-439e5b54-130e-4343-aee5-2a2ccb98d268 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] No waiting events found dispatching network-vif-plugged-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.650602] env[62235]: WARNING nova.compute.manager [req-1bd3e418-19ca-4cd4-b171-ec4f33ca22a4 req-439e5b54-130e-4343-aee5-2a2ccb98d268 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received unexpected event network-vif-plugged-64d54817-98a8-4e2e-b317-957980c86633 for instance with vm_state building and task_state spawning. [ 915.664225] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a2f8b6-e8a6-4023-b3e0-258aaa0df848 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.669450] env[62235]: DEBUG nova.network.neutron [-] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.674322] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b4e9eb-1908-59f1-d25e-cf5a1f7fd10c, 'name': SearchDatastore_Task, 'duration_secs': 0.017758} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.674952] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.675510] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 062bbe32-3edd-4b06-b7d9-882e798b3a7f/062bbe32-3edd-4b06-b7d9-882e798b3a7f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.677519] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d6d8bbf-e1db-4867-b598-4e3a979e20f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.691046] env[62235]: DEBUG nova.compute.provider_tree [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.697920] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 915.697920] env[62235]: value = "task-1271910" [ 915.697920] env[62235]: _type = "Task" [ 915.697920] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.711773] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271910, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.729342] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cb3fa8-c92a-2ac8-4095-6c5177b3c382/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.730782] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1776e7-a1f7-4535-9d8d-969c8bcce762 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.737408] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cb3fa8-c92a-2ac8-4095-6c5177b3c382/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.737408] env[62235]: ERROR oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cb3fa8-c92a-2ac8-4095-6c5177b3c382/disk-0.vmdk due to incomplete transfer. [ 915.737954] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e2f65bbc-d36b-4eca-8f51-094656111dc2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.745610] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c31453d6-3e00-4d51-bb04-3db043ecfaf3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.747644] env[62235]: DEBUG oslo_vmware.rw_handles [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cb3fa8-c92a-2ac8-4095-6c5177b3c382/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 915.747644] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Uploaded image be3e2592-5366-4a7a-b8e7-a28c16a9ddf3 to the Glance image server {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 915.749658] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 915.750931] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-89f77ad9-44d6-496b-9013-ab17f6b4796c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.752946] env[62235]: DEBUG nova.network.neutron [-] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.757780] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84d262d-f935-4407-9c14-0814970fc5d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.773440] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 915.773440] env[62235]: value = "task-1271911" [ 915.773440] env[62235]: _type = "Task" [ 915.773440] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.780324] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271911, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.788935] env[62235]: DEBUG nova.compute.manager [req-affb12b1-42e3-4d64-bc99-5dd22d14630a req-b9dff08f-cf52-43ab-a278-aeb04ae18114 service nova] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Detach interface failed, port_id=347663dc-4c1e-4e45-ab26-82dc53ace862, reason: Instance b5e4c6b2-4583-4a20-b570-99699063e172 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 915.838707] env[62235]: DEBUG nova.network.neutron [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Successfully updated port: 64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.863016] env[62235]: DEBUG oslo_concurrency.lockutils [None req-12614898-acaf-481a-ac50-35689fda9648 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.181s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.014774] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.014774] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528cdff1-b214-7021-9049-33bb4a6c50e7" [ 916.014774] env[62235]: _type = "HttpNfcLease" [ 916.014774] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.177646] env[62235]: INFO nova.compute.manager [-] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Took 1.50 seconds to deallocate network for instance. [ 916.193914] env[62235]: DEBUG nova.scheduler.client.report [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.214445] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271910, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.255926] env[62235]: INFO nova.compute.manager [-] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Took 1.57 seconds to deallocate network for instance. [ 916.287113] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271911, 'name': Destroy_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.326164] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "ecb104db-2172-4b09-867a-7a2813e03fa5" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.326645] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.326981] env[62235]: INFO nova.compute.manager [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Rebooting instance [ 916.342179] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.342399] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.342605] env[62235]: DEBUG nova.network.neutron [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.514364] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.514364] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528cdff1-b214-7021-9049-33bb4a6c50e7" [ 916.514364] env[62235]: _type = "HttpNfcLease" [ 916.514364] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 916.514673] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 916.514673] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528cdff1-b214-7021-9049-33bb4a6c50e7" [ 916.514673] env[62235]: _type = "HttpNfcLease" [ 916.514673] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 916.515375] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f0afd1-2e9d-4d94-ba0b-6fdb2eac713f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.522705] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52600f88-bb51-1815-2aaa-a102e45f9c1a/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 916.522883] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52600f88-bb51-1815-2aaa-a102e45f9c1a/disk-0.vmdk. {{(pid=62235) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 916.586964] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a34a351b-ce9f-4241-bb8b-061e6a1aeb01 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.688698] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.706137] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.850s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.713199] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.759s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.714981] env[62235]: INFO nova.compute.claims [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 916.717748] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271910, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.736828] env[62235]: INFO nova.scheduler.client.report [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Deleted allocations for instance 507fa559-1889-4f6e-81ee-93eb14d134b2 [ 916.764983] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.784877] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271911, 'name': Destroy_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.851024] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.851267] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquired lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.851455] env[62235]: DEBUG nova.network.neutron [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.900240] env[62235]: DEBUG nova.network.neutron [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.210418] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271910, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.251020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aba88733-1992-4e1a-ad59-b8885dac4dfc tempest-ImagesOneServerTestJSON-432952549 tempest-ImagesOneServerTestJSON-432952549-project-member] Lock "507fa559-1889-4f6e-81ee-93eb14d134b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.961s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.291180] env[62235]: DEBUG nova.compute.manager [req-2d5d44ed-b7c4-404e-b6a2-460802747314 req-ee2ed2d4-2859-41fb-8300-a2f6aea7981c service nova] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Received event network-vif-deleted-3c9265cd-f975-407d-8f23-ab587b57e089 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.299180] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271911, 'name': Destroy_Task, 'duration_secs': 1.488173} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.299841] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Destroyed the VM [ 917.300164] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 917.300488] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-666ec23c-a747-4018-8cc1-2ee8ef7636de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.308242] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 917.308242] env[62235]: value = "task-1271912" [ 917.308242] env[62235]: _type = "Task" [ 917.308242] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.319604] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271912, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.464797] env[62235]: DEBUG nova.network.neutron [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.711213] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271910, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.585107} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.713661] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Completed reading data from the image iterator. {{(pid=62235) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 917.713897] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52600f88-bb51-1815-2aaa-a102e45f9c1a/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 917.714343] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 062bbe32-3edd-4b06-b7d9-882e798b3a7f/062bbe32-3edd-4b06-b7d9-882e798b3a7f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.714603] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.716061] env[62235]: DEBUG nova.compute.manager [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-changed-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.716307] env[62235]: DEBUG nova.compute.manager [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing instance network info cache due to event network-changed-64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.716515] env[62235]: DEBUG oslo_concurrency.lockutils [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.717850] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805bbad0-8b22-4f8a-9143-fa4053090f87 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.721845] env[62235]: DEBUG nova.network.neutron [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Updating instance_info_cache with network_info: [{"id": "e66e8da7-6e73-460a-8812-12398d18bd9c", "address": "fa:16:3e:d6:61:e0", "network": {"id": "095f993a-0f79-4100-ab40-90d159a2bacf", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1486577542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "baccc8328255412693198d697f82cdf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ef6889-a40c-40f5-a6e5-d8726606296a", "external-id": "nsx-vlan-transportzone-537", "segmentation_id": 537, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape66e8da7-6e", "ovs_interfaceid": "e66e8da7-6e73-460a-8812-12398d18bd9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.723333] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7452e99e-5b6a-4dae-ba73-8433ebf293e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.733778] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52600f88-bb51-1815-2aaa-a102e45f9c1a/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 917.734140] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52600f88-bb51-1815-2aaa-a102e45f9c1a/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 917.737151] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-2ed1d3a9-8fa5-4a18-9d1e-64dce2f97594 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.737372] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 917.737372] env[62235]: value = "task-1271913" [ 917.737372] env[62235]: _type = "Task" [ 917.737372] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.747911] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.820242] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271912, 'name': RemoveSnapshot_Task, 'duration_secs': 0.424378} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.820524] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 917.820801] env[62235]: DEBUG nova.compute.manager [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.822470] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4585f9d-cd53-4e34-87cd-51dc69acf253 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.897157] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cbf1f5-e190-45b9-a846-bbee117b4d02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.909658] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6691d561-384f-40ac-b314-7746f5419911 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.947026] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3030f092-ee5b-48f7-b306-a98647563431 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.950027] env[62235]: DEBUG oslo_vmware.rw_handles [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52600f88-bb51-1815-2aaa-a102e45f9c1a/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 917.950322] env[62235]: INFO nova.virt.vmwareapi.images [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Downloaded image file data 0b74899a-987d-4936-af42-5d58c7e24984 [ 917.952028] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49859611-e29b-4d1f-aad5-15636bd577ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.970566] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d1d889-483c-4c6d-8394-b8bcee6563d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.975046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.976263] env[62235]: DEBUG nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Instance network_info: |[{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.976263] env[62235]: DEBUG oslo_concurrency.lockutils [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.976263] env[62235]: DEBUG nova.network.neutron [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing network info cache for port 64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.977566] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:0b:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '64d54817-98a8-4e2e-b317-957980c86633', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.985114] env[62235]: DEBUG oslo.service.loopingcall [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.985114] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee2bb708-bb66-4d79-ba3e-88bdd914b5f7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.988091] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.988590] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fea3f08b-f3ec-48bc-8767-6990f35c301d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.014023] env[62235]: DEBUG nova.compute.provider_tree [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.022065] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.022065] env[62235]: value = "task-1271915" [ 918.022065] env[62235]: _type = "Task" [ 918.022065] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.029746] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271915, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.032904] env[62235]: INFO nova.virt.vmwareapi.images [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] The imported VM was unregistered [ 918.035041] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Caching image {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 918.035279] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating directory with path [datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.035799] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d624a71-e122-4f11-8de2-a4c487439489 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.050227] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created directory with path [datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.050449] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc/OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc.vmdk to [datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984/0b74899a-987d-4936-af42-5d58c7e24984.vmdk. {{(pid=62235) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 918.050725] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-75325574-9f22-402f-85cf-09eacca1aee5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.056851] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 918.056851] env[62235]: value = "task-1271916" [ 918.056851] env[62235]: _type = "Task" [ 918.056851] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.065059] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271916, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.230405] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Releasing lock "refresh_cache-ecb104db-2172-4b09-867a-7a2813e03fa5" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.232959] env[62235]: DEBUG nova.compute.manager [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.233738] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a0755f-cd15-46c9-b968-e27790d51741 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.253023] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116914} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.253023] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.253023] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c05d6b-9813-4332-b2ac-0ab5f4212c13 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.275705] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 062bbe32-3edd-4b06-b7d9-882e798b3a7f/062bbe32-3edd-4b06-b7d9-882e798b3a7f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.276140] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b12d4e19-1bc3-46db-b468-5b95eaedcd71 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.297818] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 918.297818] env[62235]: value = "task-1271917" [ 918.297818] env[62235]: _type = "Task" [ 918.297818] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.306290] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271917, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.336900] env[62235]: INFO nova.compute.manager [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Shelve offloading [ 918.338886] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 918.339206] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6551871-0910-4813-99f0-325fee91fc94 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.345765] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 918.345765] env[62235]: value = "task-1271918" [ 918.345765] env[62235]: _type = "Task" [ 918.345765] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.357377] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 918.357626] env[62235]: DEBUG nova.compute.manager [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.358463] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d65e51-2f44-4102-b212-1f14a3693952 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.364873] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.365092] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.365291] env[62235]: DEBUG nova.network.neutron [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.516815] env[62235]: DEBUG nova.scheduler.client.report [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.529856] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271915, 'name': CreateVM_Task, 'duration_secs': 0.36713} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.530317] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.531464] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.531602] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.531929] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.533171] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6157ff0-01c2-4e54-9c04-0a6f061a679f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.538278] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 918.538278] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b65412-9ada-5881-f42e-259e048cb4b3" [ 918.538278] env[62235]: _type = "Task" [ 918.538278] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.545841] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b65412-9ada-5881-f42e-259e048cb4b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.571021] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271916, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.808744] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.912343] env[62235]: DEBUG nova.network.neutron [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updated VIF entry in instance network info cache for port 64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.912736] env[62235]: DEBUG nova.network.neutron [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.022537] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.023109] env[62235]: DEBUG nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.029036] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.341s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.029325] env[62235]: DEBUG nova.objects.instance [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lazy-loading 'resources' on Instance uuid b5e4c6b2-4583-4a20-b570-99699063e172 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.051246] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b65412-9ada-5881-f42e-259e048cb4b3, 'name': SearchDatastore_Task, 'duration_secs': 0.028234} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.051246] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.051544] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.051544] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.051804] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.052118] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.052443] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74ffc50a-8d4e-433e-b232-5162d7c09d0b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.067931] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271916, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.069329] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.069546] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.072391] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdc872be-7e5c-4618-af99-108590abeea4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.083165] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 919.083165] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522cc6b8-4596-f9bd-d67d-9e94cd454341" [ 919.083165] env[62235]: _type = "Task" [ 919.083165] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.092451] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522cc6b8-4596-f9bd-d67d-9e94cd454341, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.116782] env[62235]: DEBUG nova.network.neutron [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Updating instance_info_cache with network_info: [{"id": "50f22daa-f003-4786-a006-57fc3b25326d", "address": "fa:16:3e:c5:de:67", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50f22daa-f0", "ovs_interfaceid": "50f22daa-f003-4786-a006-57fc3b25326d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.254063] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5ffc3d-cd91-4ade-be9d-d5652c4d92a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.263045] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Doing hard reboot of VM {{(pid=62235) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 919.263681] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-fe6aef45-71e9-43b5-af2a-390ac7ac9a1d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.271047] env[62235]: DEBUG oslo_vmware.api [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 919.271047] env[62235]: value = "task-1271919" [ 919.271047] env[62235]: _type = "Task" [ 919.271047] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.286874] env[62235]: DEBUG oslo_vmware.api [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271919, 'name': ResetVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.310343] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271917, 'name': ReconfigVM_Task, 'duration_secs': 0.5996} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.310676] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 062bbe32-3edd-4b06-b7d9-882e798b3a7f/062bbe32-3edd-4b06-b7d9-882e798b3a7f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.311547] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf44b091-9eb9-4918-a20a-b3c90a007b15 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.318798] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 919.318798] env[62235]: value = "task-1271920" [ 919.318798] env[62235]: _type = "Task" [ 919.318798] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.328079] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271920, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.415655] env[62235]: DEBUG oslo_concurrency.lockutils [req-b4a4a2f6-547a-4698-ab66-309be4a0794a req-f6944884-b045-4ca0-9592-5f959ebe41d2 service nova] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.528910] env[62235]: DEBUG nova.compute.utils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 919.530655] env[62235]: DEBUG nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 919.530959] env[62235]: DEBUG nova.network.neutron [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 919.568623] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271916, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.594059] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522cc6b8-4596-f9bd-d67d-9e94cd454341, 'name': SearchDatastore_Task, 'duration_secs': 0.086687} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.595175] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c953f204-d4d4-498e-b23b-c769413f2bd7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.600759] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 919.600759] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5224d8f5-75a8-917e-0ed2-11f43de877d1" [ 919.600759] env[62235]: _type = "Task" [ 919.600759] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.606255] env[62235]: DEBUG nova.policy [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b64f6d6696c042079cced4c83d161f6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4469550d32b4482a9d5e9244cbd681e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 919.615060] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5224d8f5-75a8-917e-0ed2-11f43de877d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.619689] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.706963] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3bfc6a-db69-4a7f-bd12-0b241aba3e8c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.716748] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89df7f64-9c61-47ca-96c1-080438a1d853 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.755056] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4454dcee-1ce9-4386-a6bb-050054d6a9ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.764855] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44bfda57-ea82-47ef-b2c1-063bd23bb500 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.785569] env[62235]: DEBUG nova.compute.provider_tree [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.799614] env[62235]: DEBUG oslo_vmware.api [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271919, 'name': ResetVM_Task, 'duration_secs': 0.110703} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.800780] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Did hard reboot of VM {{(pid=62235) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 919.801056] env[62235]: DEBUG nova.compute.manager [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.801880] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dfb4dd-b35d-4f13-805d-28664d8fce0e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.833057] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271920, 'name': Rename_Task, 'duration_secs': 0.161765} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.834071] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.834071] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1d49fa2-a2d3-4bc0-bd90-f389e330e3c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.842213] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 919.842213] env[62235]: value = "task-1271921" [ 919.842213] env[62235]: _type = "Task" [ 919.842213] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.853688] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271921, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.034512] env[62235]: DEBUG nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.059694] env[62235]: DEBUG nova.network.neutron [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Successfully created port: e3865625-b03d-42d2-8ba3-8829b7996727 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.070752] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271916, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.099946] env[62235]: DEBUG nova.compute.manager [req-1920d06b-7d75-4c8f-9106-cfbeba02fb8d req-ce465a9b-4cc4-4fc4-96cf-67cda37c78f1 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Received event network-vif-unplugged-50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.100619] env[62235]: DEBUG oslo_concurrency.lockutils [req-1920d06b-7d75-4c8f-9106-cfbeba02fb8d req-ce465a9b-4cc4-4fc4-96cf-67cda37c78f1 service nova] Acquiring lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.101357] env[62235]: DEBUG oslo_concurrency.lockutils [req-1920d06b-7d75-4c8f-9106-cfbeba02fb8d req-ce465a9b-4cc4-4fc4-96cf-67cda37c78f1 service nova] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.101357] env[62235]: DEBUG oslo_concurrency.lockutils [req-1920d06b-7d75-4c8f-9106-cfbeba02fb8d req-ce465a9b-4cc4-4fc4-96cf-67cda37c78f1 service nova] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.101535] env[62235]: DEBUG nova.compute.manager [req-1920d06b-7d75-4c8f-9106-cfbeba02fb8d req-ce465a9b-4cc4-4fc4-96cf-67cda37c78f1 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] No waiting events found dispatching network-vif-unplugged-50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.101758] env[62235]: WARNING nova.compute.manager [req-1920d06b-7d75-4c8f-9106-cfbeba02fb8d req-ce465a9b-4cc4-4fc4-96cf-67cda37c78f1 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Received unexpected event network-vif-unplugged-50f22daa-f003-4786-a006-57fc3b25326d for instance with vm_state shelved and task_state shelving_offloading. [ 920.113489] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5224d8f5-75a8-917e-0ed2-11f43de877d1, 'name': SearchDatastore_Task, 'duration_secs': 0.087398} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.114855] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.114855] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 2068403d-5870-4daa-b4d6-acaf39209488/2068403d-5870-4daa-b4d6-acaf39209488.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.114855] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1e35399-2e07-47f2-a14b-505c62596ade {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.121996] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 920.121996] env[62235]: value = "task-1271922" [ 920.121996] env[62235]: _type = "Task" [ 920.121996] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.128522] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.129091] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c138b55-fd0e-44df-a33b-a43870d11332 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.135637] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.140515] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 920.140854] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bd6bd16-6d2e-426e-8803-a05afa727b38 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.294964] env[62235]: DEBUG nova.scheduler.client.report [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.316971] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c770b8cc-c492-47e6-9779-38a4e6a7a93a tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.989s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.357116] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271921, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.574459] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271916, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.636841] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.811622] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.814163] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.050s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.814397] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.837877] env[62235]: INFO nova.scheduler.client.report [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Deleted allocations for instance b5e4c6b2-4583-4a20-b570-99699063e172 [ 920.840032] env[62235]: INFO nova.scheduler.client.report [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted allocations for instance 44ed8739-ad20-41a3-abe1-38176c1bee25 [ 920.857284] env[62235]: DEBUG oslo_vmware.api [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1271921, 'name': PowerOnVM_Task, 'duration_secs': 0.859387} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.857509] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.857738] env[62235]: INFO nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Took 10.44 seconds to spawn the instance on the hypervisor. [ 920.858009] env[62235]: DEBUG nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.858947] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2997d760-9dbc-45a9-88dc-be3ff6073d83 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.045157] env[62235]: DEBUG nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.078014] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271916, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.889708} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.081185] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.081561] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.081774] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.082013] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.082313] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.082489] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.082750] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.082971] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.084400] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.084563] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.084888] env[62235]: DEBUG nova.virt.hardware [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.085202] env[62235]: INFO nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc/OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc.vmdk to [datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984/0b74899a-987d-4936-af42-5d58c7e24984.vmdk. [ 921.085449] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Cleaning up location [datastore1] OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 921.085625] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_d77be99e-c71d-492f-a9a1-9cdead5328fc {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.086569] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149f220f-3d0b-4234-b620-43e32599d2d9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.089784] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3dcf0374-909b-420d-9f29-cff55ed35036 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.099539] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9d6ce4-07e9-47ea-832f-28bf40886fd4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.107336] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 921.107336] env[62235]: value = "task-1271924" [ 921.107336] env[62235]: _type = "Task" [ 921.107336] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.132377] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271924, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.139247] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271922, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.352285] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f9d82ad2-7a4b-4f4f-8571-640fced9e758 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "44ed8739-ad20-41a3-abe1-38176c1bee25" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.955s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.353730] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f57a0fcc-7a23-4183-98b1-21de0e3fd575 tempest-ImagesTestJSON-2144061415 tempest-ImagesTestJSON-2144061415-project-member] Lock "b5e4c6b2-4583-4a20-b570-99699063e172" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.344s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.380232] env[62235]: INFO nova.compute.manager [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Took 15.27 seconds to build instance. [ 921.617352] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "ecb104db-2172-4b09-867a-7a2813e03fa5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.617609] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.617870] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "ecb104db-2172-4b09-867a-7a2813e03fa5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.618016] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.618213] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.622284] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271924, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190924} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.622284] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.622284] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984/0b74899a-987d-4936-af42-5d58c7e24984.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.622284] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984/0b74899a-987d-4936-af42-5d58c7e24984.vmdk to [datastore1] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.622284] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76acf31f-e828-470b-90c5-b26f6b57f898 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.623891] env[62235]: INFO nova.compute.manager [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Terminating instance [ 921.632260] env[62235]: DEBUG nova.compute.manager [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 921.632473] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.632812] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 921.632812] env[62235]: value = "task-1271925" [ 921.632812] env[62235]: _type = "Task" [ 921.632812] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.633545] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe83d6f3-5452-43e2-a90a-06fc005455f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.643425] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271922, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.209414} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.646104] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 2068403d-5870-4daa-b4d6-acaf39209488/2068403d-5870-4daa-b4d6-acaf39209488.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.646104] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.646396] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.646611] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-278bcd6c-d1ed-4ed4-8616-79060152eb14 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.648478] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-744f23ad-3c89-425f-8f6e-1d191c6d2bcf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.653203] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271925, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.657044] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 921.657044] env[62235]: value = "task-1271926" [ 921.657044] env[62235]: _type = "Task" [ 921.657044] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.660795] env[62235]: DEBUG oslo_vmware.api [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 921.660795] env[62235]: value = "task-1271927" [ 921.660795] env[62235]: _type = "Task" [ 921.660795] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.667201] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271926, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.671858] env[62235]: DEBUG oslo_vmware.api [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.780148] env[62235]: DEBUG nova.compute.manager [req-4e4a638a-816b-4dce-8871-939a325d4e8c req-807f41e8-92bd-4ec4-b121-436a23a0c456 service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Received event network-vif-plugged-e3865625-b03d-42d2-8ba3-8829b7996727 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.780536] env[62235]: DEBUG oslo_concurrency.lockutils [req-4e4a638a-816b-4dce-8871-939a325d4e8c req-807f41e8-92bd-4ec4-b121-436a23a0c456 service nova] Acquiring lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.780607] env[62235]: DEBUG oslo_concurrency.lockutils [req-4e4a638a-816b-4dce-8871-939a325d4e8c req-807f41e8-92bd-4ec4-b121-436a23a0c456 service nova] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.780892] env[62235]: DEBUG oslo_concurrency.lockutils [req-4e4a638a-816b-4dce-8871-939a325d4e8c req-807f41e8-92bd-4ec4-b121-436a23a0c456 service nova] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.780990] env[62235]: DEBUG nova.compute.manager [req-4e4a638a-816b-4dce-8871-939a325d4e8c req-807f41e8-92bd-4ec4-b121-436a23a0c456 service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] No waiting events found dispatching network-vif-plugged-e3865625-b03d-42d2-8ba3-8829b7996727 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.781211] env[62235]: WARNING nova.compute.manager [req-4e4a638a-816b-4dce-8871-939a325d4e8c req-807f41e8-92bd-4ec4-b121-436a23a0c456 service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Received unexpected event network-vif-plugged-e3865625-b03d-42d2-8ba3-8829b7996727 for instance with vm_state building and task_state spawning. [ 921.882294] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6f0d11eb-488f-4973-b585-05974d3ccd08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.776s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.983136] env[62235]: DEBUG nova.network.neutron [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Successfully updated port: e3865625-b03d-42d2-8ba3-8829b7996727 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.145725] env[62235]: DEBUG nova.compute.manager [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Received event network-changed-50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.146502] env[62235]: DEBUG nova.compute.manager [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Refreshing instance network info cache due to event network-changed-50f22daa-f003-4786-a006-57fc3b25326d. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 922.146502] env[62235]: DEBUG oslo_concurrency.lockutils [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] Acquiring lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.146502] env[62235]: DEBUG oslo_concurrency.lockutils [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] Acquired lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.146760] env[62235]: DEBUG nova.network.neutron [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Refreshing network info cache for port 50f22daa-f003-4786-a006-57fc3b25326d {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.155082] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271925, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.173892] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271926, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12053} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.177510] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.178333] env[62235]: DEBUG oslo_vmware.api [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271927, 'name': PowerOffVM_Task, 'duration_secs': 0.21301} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.179193] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f5a0b3-b21a-44cc-8526-e335dbf7f939 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.182631] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.182885] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.183282] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b186363-81b0-4122-a52b-5bae8dd62737 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.208020] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 2068403d-5870-4daa-b4d6-acaf39209488/2068403d-5870-4daa-b4d6-acaf39209488.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.208512] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f86c6af-7a9a-42f9-801a-dd8002026b05 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.231779] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 922.231779] env[62235]: value = "task-1271929" [ 922.231779] env[62235]: _type = "Task" [ 922.231779] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.241909] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271929, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.261416] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.261865] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.262195] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Deleting the datastore file [datastore2] ecb104db-2172-4b09-867a-7a2813e03fa5 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.262578] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9492a403-f416-4d45-8010-4c933e334018 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.272463] env[62235]: DEBUG oslo_vmware.api [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for the task: (returnval){ [ 922.272463] env[62235]: value = "task-1271930" [ 922.272463] env[62235]: _type = "Task" [ 922.272463] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.287085] env[62235]: DEBUG oslo_vmware.api [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271930, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.485838] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.485996] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.486161] env[62235]: DEBUG nova.network.neutron [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.648484] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271925, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.750083] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271929, 'name': ReconfigVM_Task, 'duration_secs': 0.314356} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.755116] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 2068403d-5870-4daa-b4d6-acaf39209488/2068403d-5870-4daa-b4d6-acaf39209488.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.756151] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d44b121c-8c14-4696-9b56-ed1a6954354e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.766194] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 922.766194] env[62235]: value = "task-1271931" [ 922.766194] env[62235]: _type = "Task" [ 922.766194] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.776726] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271931, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.786924] env[62235]: DEBUG oslo_vmware.api [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Task: {'id': task-1271930, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.331107} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.787098] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.787341] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.787485] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.787862] env[62235]: INFO nova.compute.manager [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 922.787979] env[62235]: DEBUG oslo.service.loopingcall [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.788205] env[62235]: DEBUG nova.compute.manager [-] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.788297] env[62235]: DEBUG nova.network.neutron [-] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.947883] env[62235]: DEBUG nova.network.neutron [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Updated VIF entry in instance network info cache for port 50f22daa-f003-4786-a006-57fc3b25326d. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.948579] env[62235]: DEBUG nova.network.neutron [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Updating instance_info_cache with network_info: [{"id": "50f22daa-f003-4786-a006-57fc3b25326d", "address": "fa:16:3e:c5:de:67", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": null, "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap50f22daa-f0", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.031887] env[62235]: DEBUG nova.network.neutron [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.148360] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271925, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.255427] env[62235]: DEBUG nova.network.neutron [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance_info_cache with network_info: [{"id": "e3865625-b03d-42d2-8ba3-8829b7996727", "address": "fa:16:3e:2f:a8:fc", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3865625-b0", "ovs_interfaceid": "e3865625-b03d-42d2-8ba3-8829b7996727", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.290374] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271931, 'name': Rename_Task, 'duration_secs': 0.15301} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.290749] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.291085] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-263d7b34-d540-4242-8e85-3154ee8fb459 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.298917] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 923.298917] env[62235]: value = "task-1271932" [ 923.298917] env[62235]: _type = "Task" [ 923.298917] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.307936] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271932, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.453353] env[62235]: DEBUG oslo_concurrency.lockutils [req-d52d787e-cd82-4fd0-bd20-62d5ce6a0d59 req-65ad3655-6415-4346-866e-9b9f8cedff86 service nova] Releasing lock "refresh_cache-d98ab48f-c325-473a-8ee8-57ecd3b81c8b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.654106] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271925, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.759065] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.761103] env[62235]: DEBUG nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Instance network_info: |[{"id": "e3865625-b03d-42d2-8ba3-8829b7996727", "address": "fa:16:3e:2f:a8:fc", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3865625-b0", "ovs_interfaceid": "e3865625-b03d-42d2-8ba3-8829b7996727", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.761103] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:a8:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3865625-b03d-42d2-8ba3-8829b7996727', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.772183] env[62235]: DEBUG oslo.service.loopingcall [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.772183] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.772183] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6a93953-e126-40e9-b51e-dde65ed6df54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.800542] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.800542] env[62235]: value = "task-1271933" [ 923.800542] env[62235]: _type = "Task" [ 923.800542] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.827015] env[62235]: DEBUG nova.compute.manager [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Received event network-changed-e3865625-b03d-42d2-8ba3-8829b7996727 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.827015] env[62235]: DEBUG nova.compute.manager [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Refreshing instance network info cache due to event network-changed-e3865625-b03d-42d2-8ba3-8829b7996727. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.827015] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] Acquiring lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.827015] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] Acquired lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.827015] env[62235]: DEBUG nova.network.neutron [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Refreshing network info cache for port e3865625-b03d-42d2-8ba3-8829b7996727 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.827214] env[62235]: DEBUG oslo_vmware.api [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271932, 'name': PowerOnVM_Task, 'duration_secs': 0.501027} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.827374] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271933, 'name': CreateVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.827638] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.827838] env[62235]: INFO nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Took 8.80 seconds to spawn the instance on the hypervisor. [ 923.828111] env[62235]: DEBUG nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.829117] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a662d89-e949-4ffe-a057-d1a9460accde {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.875888] env[62235]: DEBUG nova.network.neutron [-] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.150821] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271925, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.176678] env[62235]: DEBUG nova.compute.manager [req-d07b4474-5da6-4b99-8d39-1f126045e559 req-6abae098-3042-4aee-8330-b482890cc9fa service nova] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Received event network-vif-deleted-e66e8da7-6e73-460a-8812-12398d18bd9c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.188272] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.188272] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.188272] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleting the datastore file [datastore1] d98ab48f-c325-473a-8ee8-57ecd3b81c8b {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.188517] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d402a9d8-7114-4a98-a8b2-2ad61ccb0f93 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.194806] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 924.194806] env[62235]: value = "task-1271934" [ 924.194806] env[62235]: _type = "Task" [ 924.194806] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.204817] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271934, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.321414] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271933, 'name': CreateVM_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.350071] env[62235]: INFO nova.compute.manager [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Took 17.99 seconds to build instance. [ 924.382342] env[62235]: INFO nova.compute.manager [-] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Took 1.59 seconds to deallocate network for instance. [ 924.566978] env[62235]: DEBUG nova.network.neutron [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updated VIF entry in instance network info cache for port e3865625-b03d-42d2-8ba3-8829b7996727. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.567378] env[62235]: DEBUG nova.network.neutron [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance_info_cache with network_info: [{"id": "e3865625-b03d-42d2-8ba3-8829b7996727", "address": "fa:16:3e:2f:a8:fc", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3865625-b0", "ovs_interfaceid": "e3865625-b03d-42d2-8ba3-8829b7996727", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.657422] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271925, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.610323} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.657701] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0b74899a-987d-4936-af42-5d58c7e24984/0b74899a-987d-4936-af42-5d58c7e24984.vmdk to [datastore1] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.658569] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1654d4ab-c9d7-4e4a-a75e-4f319bb9d28f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.683933] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.684283] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b3c5e21-198b-44fe-8b5f-5ec10ac9899e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.716945] env[62235]: DEBUG oslo_vmware.api [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1271934, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195831} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.717272] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 924.717272] env[62235]: value = "task-1271935" [ 924.717272] env[62235]: _type = "Task" [ 924.717272] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.717501] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.717743] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.717856] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.729533] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271935, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.747199] env[62235]: INFO nova.scheduler.client.report [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted allocations for instance d98ab48f-c325-473a-8ee8-57ecd3b81c8b [ 924.815356] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271933, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.854284] env[62235]: DEBUG oslo_concurrency.lockutils [None req-98d60a6c-7009-4904-ab2c-102d76aa7896 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "2068403d-5870-4daa-b4d6-acaf39209488" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.501s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.890800] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.890982] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.891222] env[62235]: DEBUG nova.objects.instance [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lazy-loading 'resources' on Instance uuid ecb104db-2172-4b09-867a-7a2813e03fa5 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.072574] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] Releasing lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.072574] env[62235]: DEBUG nova.compute.manager [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Received event network-changed-984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.072574] env[62235]: DEBUG nova.compute.manager [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Refreshing instance network info cache due to event network-changed-984bbc0d-59a6-46fc-bb73-bcc8454e14e3. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.072574] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] Acquiring lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.072574] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] Acquired lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.072574] env[62235]: DEBUG nova.network.neutron [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Refreshing network info cache for port 984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 925.231890] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271935, 'name': ReconfigVM_Task, 'duration_secs': 0.287171} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.232570] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 0e9182f8-81e1-4ca2-97b6-0066faba0b22/0e9182f8-81e1-4ca2-97b6-0066faba0b22.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.234117] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encryption_options': None, 'encryption_format': None, 'device_type': 'disk', 'encrypted': False, 'disk_bus': None, 'size': 0, 'encryption_secret_uuid': None, 'boot_index': 0, 'device_name': '/dev/sda', 'image_id': '02c64327-1eca-4d55-9f01-62fa8b2a2334'}], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sdb', 'attachment_id': 'a4d7f10b-b6f8-4e4b-bc40-2ac72b5d7820', 'device_type': None, 'disk_bus': None, 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273467', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'name': 'volume-ff64a51b-f62a-4682-83fb-4e26eae16346', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0e9182f8-81e1-4ca2-97b6-0066faba0b22', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'serial': 'ff64a51b-f62a-4682-83fb-4e26eae16346'}, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=62235) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 925.235517] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 925.235517] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273467', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'name': 'volume-ff64a51b-f62a-4682-83fb-4e26eae16346', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0e9182f8-81e1-4ca2-97b6-0066faba0b22', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'serial': 'ff64a51b-f62a-4682-83fb-4e26eae16346'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 925.235901] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b82984-7f1e-4515-9cc0-dec1471f1f7b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.258994] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.261508] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333b72cc-598d-47f2-a9c2-aafa4ffaa553 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.296016] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] volume-ff64a51b-f62a-4682-83fb-4e26eae16346/volume-ff64a51b-f62a-4682-83fb-4e26eae16346.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.296016] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18c30e93-f40a-41ed-8578-f977148c71cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.317433] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271933, 'name': CreateVM_Task, 'duration_secs': 1.039409} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.322023] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.322023] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 925.322023] env[62235]: value = "task-1271936" [ 925.322023] env[62235]: _type = "Task" [ 925.322023] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.322023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.322023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.322023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.322023] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16598c88-2206-415a-b04b-1e37966b6820 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.329531] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271936, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.330882] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 925.330882] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52480010-d9ad-9274-879a-fae58a1d5774" [ 925.330882] env[62235]: _type = "Task" [ 925.330882] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.348387] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52480010-d9ad-9274-879a-fae58a1d5774, 'name': SearchDatastore_Task, 'duration_secs': 0.012876} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.348687] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.349218] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.349218] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.349349] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.349515] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.349773] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f27b344-8239-4ead-b6b9-46b9fd31df75 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.358436] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.358558] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.359301] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de7682e4-db17-48ef-9238-ed9595eb4d33 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.364702] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 925.364702] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521eec44-7e14-4fb6-81a8-f8ab4d7eb8ab" [ 925.364702] env[62235]: _type = "Task" [ 925.364702] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.373046] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521eec44-7e14-4fb6-81a8-f8ab4d7eb8ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.518828] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8630ee5c-acf3-4f37-999c-bb307f56fea1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.526640] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50f7de2-657a-4021-ac47-acde34c57488 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.562218] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0abcaf-805b-46ae-936c-7c1055374de7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.571238] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c17805-f580-4543-a07b-914c49f60aa3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.588555] env[62235]: DEBUG nova.compute.provider_tree [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.831173] env[62235]: DEBUG nova.network.neutron [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Updated VIF entry in instance network info cache for port 984bbc0d-59a6-46fc-bb73-bcc8454e14e3. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 925.831812] env[62235]: DEBUG nova.network.neutron [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Updating instance_info_cache with network_info: [{"id": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "address": "fa:16:3e:fa:ee:90", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984bbc0d-59", "ovs_interfaceid": "984bbc0d-59a6-46fc-bb73-bcc8454e14e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.836294] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271936, 'name': ReconfigVM_Task, 'duration_secs': 0.297935} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.836799] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfigured VM instance instance-0000003d to attach disk [datastore1] volume-ff64a51b-f62a-4682-83fb-4e26eae16346/volume-ff64a51b-f62a-4682-83fb-4e26eae16346.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.843118] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.844781] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa3dd28a-896d-4e32-969b-d5dd2d2ba390 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.860735] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 925.860735] env[62235]: value = "task-1271937" [ 925.860735] env[62235]: _type = "Task" [ 925.860735] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.877375] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271937, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.881572] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521eec44-7e14-4fb6-81a8-f8ab4d7eb8ab, 'name': SearchDatastore_Task, 'duration_secs': 0.012682} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.883095] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7910b53d-0ee8-4fd0-a85d-9b7ba0f52c63 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.888752] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 925.888752] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210b8b4-56dc-390f-d888-6b10f3998674" [ 925.888752] env[62235]: _type = "Task" [ 925.888752] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.898728] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210b8b4-56dc-390f-d888-6b10f3998674, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.091693] env[62235]: DEBUG nova.scheduler.client.report [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.285696] env[62235]: DEBUG nova.compute.manager [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-changed-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.285696] env[62235]: DEBUG nova.compute.manager [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing instance network info cache due to event network-changed-64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 926.285696] env[62235]: DEBUG oslo_concurrency.lockutils [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.285696] env[62235]: DEBUG oslo_concurrency.lockutils [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.285696] env[62235]: DEBUG nova.network.neutron [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing network info cache for port 64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.338223] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fe93b47-5a1a-4930-90c9-4130c9776b76 req-49400911-5687-4311-b473-4f8d99e725fe service nova] Releasing lock "refresh_cache-062bbe32-3edd-4b06-b7d9-882e798b3a7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.371241] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271937, 'name': ReconfigVM_Task, 'duration_secs': 0.174304} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.371711] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273467', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'name': 'volume-ff64a51b-f62a-4682-83fb-4e26eae16346', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0e9182f8-81e1-4ca2-97b6-0066faba0b22', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'serial': 'ff64a51b-f62a-4682-83fb-4e26eae16346'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 926.372260] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63238851-644a-41f2-be1d-8a7c4ec153f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.379802] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 926.379802] env[62235]: value = "task-1271938" [ 926.379802] env[62235]: _type = "Task" [ 926.379802] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.393883] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271938, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.402848] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210b8b4-56dc-390f-d888-6b10f3998674, 'name': SearchDatastore_Task, 'duration_secs': 0.011847} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.403243] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.403587] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 05aae40a-37ae-4bf1-91d0-02a7228fcff2/05aae40a-37ae-4bf1-91d0-02a7228fcff2.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.403940] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa3e1e9a-c5b4-46b7-a7e8-667db3841861 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.413079] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 926.413079] env[62235]: value = "task-1271939" [ 926.413079] env[62235]: _type = "Task" [ 926.413079] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.424530] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.601542] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.604972] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.346s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.606065] env[62235]: DEBUG nova.objects.instance [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'resources' on Instance uuid d98ab48f-c325-473a-8ee8-57ecd3b81c8b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.639032] env[62235]: INFO nova.scheduler.client.report [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Deleted allocations for instance ecb104db-2172-4b09-867a-7a2813e03fa5 [ 926.863025] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 926.863520] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 926.863713] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 926.892267] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271938, 'name': Rename_Task, 'duration_secs': 0.150065} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.893680] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 926.896444] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28151d46-78c0-4e10-8921-95347df3a5e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.898404] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.898634] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.904833] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 926.904833] env[62235]: value = "task-1271940" [ 926.904833] env[62235]: _type = "Task" [ 926.904833] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.914321] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.926667] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458966} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.927234] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 05aae40a-37ae-4bf1-91d0-02a7228fcff2/05aae40a-37ae-4bf1-91d0-02a7228fcff2.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.927624] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.928101] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c296542-6878-43b8-af0c-cf4bde9f8597 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.938147] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 926.938147] env[62235]: value = "task-1271941" [ 926.938147] env[62235]: _type = "Task" [ 926.938147] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.946694] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271941, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.110825] env[62235]: DEBUG nova.objects.instance [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'numa_topology' on Instance uuid d98ab48f-c325-473a-8ee8-57ecd3b81c8b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.148149] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e94eb05-b4fc-4f1f-bacd-7c863b08d179 tempest-InstanceActionsTestJSON-1005882931 tempest-InstanceActionsTestJSON-1005882931-project-member] Lock "ecb104db-2172-4b09-867a-7a2813e03fa5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.530s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.173446] env[62235]: DEBUG nova.network.neutron [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updated VIF entry in instance network info cache for port 64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.173781] env[62235]: DEBUG nova.network.neutron [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.175604] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "a3336b20-5737-4e5a-94a1-19fec140aa60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.175793] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.401308] env[62235]: DEBUG nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.419588] env[62235]: DEBUG oslo_vmware.api [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1271940, 'name': PowerOnVM_Task, 'duration_secs': 0.492526} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.420118] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 927.448866] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271941, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066249} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.449219] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.451300] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db340ae-627a-4d23-a9a3-6d6dc6d698bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.480436] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 05aae40a-37ae-4bf1-91d0-02a7228fcff2/05aae40a-37ae-4bf1-91d0-02a7228fcff2.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.481134] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3e038f2-bd2b-4a38-bd63-ff3ae3a2081c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.504023] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 927.504023] env[62235]: value = "task-1271942" [ 927.504023] env[62235]: _type = "Task" [ 927.504023] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.511833] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271942, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.540509] env[62235]: DEBUG nova.compute.manager [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.541468] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb10df61-3dca-4a74-a1e1-23d2e91711aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.613772] env[62235]: DEBUG nova.objects.base [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 927.678123] env[62235]: DEBUG oslo_concurrency.lockutils [req-1ae3fa9e-55aa-49ca-8090-b4578c11ec3f req-070859d8-811b-4bf1-8681-491c843e450e service nova] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.678614] env[62235]: DEBUG nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.742602] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9e2550-9754-41a2-86cf-38e54acd12f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.751832] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309bb586-8eb0-4562-ad9f-103e3ad94267 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.785207] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c1130a-373b-49a5-a5ec-949d0835b228 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.794047] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4415330f-ae10-4dba-8438-70d98e8562eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.806020] env[62235]: DEBUG nova.compute.provider_tree [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.873736] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.873937] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.874134] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 927.924265] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.018243] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.065911] env[62235]: DEBUG oslo_concurrency.lockutils [None req-965048f3-b791-4061-b604-8a496eb4566c tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 46.240s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.204675] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.309448] env[62235]: DEBUG nova.scheduler.client.report [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.515998] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271942, 'name': ReconfigVM_Task, 'duration_secs': 0.696955} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.516325] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 05aae40a-37ae-4bf1-91d0-02a7228fcff2/05aae40a-37ae-4bf1-91d0-02a7228fcff2.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.516959] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4f54976-f5e9-4495-8fd8-1effa8553172 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.523905] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 928.523905] env[62235]: value = "task-1271943" [ 928.523905] env[62235]: _type = "Task" [ 928.523905] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.531969] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271943, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.825058] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.217s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.828546] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.904s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.830332] env[62235]: INFO nova.compute.claims [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.035513] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271943, 'name': Rename_Task, 'duration_secs': 0.135385} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.035804] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.038875] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da1a6319-edc7-47dd-958c-09deabfb2a5d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.049270] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 929.049270] env[62235]: value = "task-1271944" [ 929.049270] env[62235]: _type = "Task" [ 929.049270] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.057772] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271944, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.337938] env[62235]: DEBUG oslo_concurrency.lockutils [None req-93c0b9f0-b7be-43cf-aef3-6dd837b6b44e tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.115s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.338900] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.496s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.339184] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.339572] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.340946] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.344067] env[62235]: INFO nova.compute.manager [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Terminating instance [ 929.349873] env[62235]: DEBUG nova.compute.manager [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 929.350150] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.352647] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11de949e-6c18-460f-a674-f625887b7d7c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.361213] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c74b217-625e-491a-9c6f-fa8b8a365e08 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.390396] env[62235]: WARNING nova.virt.vmwareapi.vmops [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d98ab48f-c325-473a-8ee8-57ecd3b81c8b could not be found. [ 929.390725] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.390986] env[62235]: INFO nova.compute.manager [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 929.391349] env[62235]: DEBUG oslo.service.loopingcall [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.394464] env[62235]: DEBUG nova.compute.manager [-] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.394621] env[62235]: DEBUG nova.network.neutron [-] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 929.487748] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "eccde807-664e-4488-bec8-2d6919c2fc91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.488151] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.498955] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [{"id": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "address": "fa:16:3e:99:60:de", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6a488c2-b3", "ovs_interfaceid": "f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.567443] env[62235]: DEBUG oslo_vmware.api [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271944, 'name': PowerOnVM_Task, 'duration_secs': 0.45816} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.567443] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.567443] env[62235]: INFO nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Took 8.52 seconds to spawn the instance on the hypervisor. [ 929.567443] env[62235]: DEBUG nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.568423] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4022336f-ad29-4087-81d8-371b36569b7a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.943211] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.943211] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.991285] env[62235]: DEBUG nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.004207] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-0e9182f8-81e1-4ca2-97b6-0066faba0b22" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.004508] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 930.004732] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.005023] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.005261] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.005870] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.006226] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.006741] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.006741] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 930.006741] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.076984] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db70eeeb-c468-4311-8602-0d30f4455d9f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.095016] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc69530e-15cc-4211-88b8-6e05f72c2303 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.099405] env[62235]: INFO nova.compute.manager [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Took 15.17 seconds to build instance. [ 930.130757] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2592c63e-e4bd-472a-9f73-00542ec80805 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.139562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8883a86e-cc59-4c6c-8e6f-1c7f80e20298 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.153556] env[62235]: DEBUG nova.compute.provider_tree [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.310217] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "79dafae1-e389-4e04-af20-577b8aac06a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.310711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "79dafae1-e389-4e04-af20-577b8aac06a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.446678] env[62235]: DEBUG nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.511953] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.517101] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.567112] env[62235]: DEBUG nova.network.neutron [-] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.601823] env[62235]: DEBUG oslo_concurrency.lockutils [None req-745c7f64-6dc2-4d2c-aba0-ed1267ad533a tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.686s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.656556] env[62235]: DEBUG nova.scheduler.client.report [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.812809] env[62235]: DEBUG nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.981046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.071752] env[62235]: INFO nova.compute.manager [-] [instance: d98ab48f-c325-473a-8ee8-57ecd3b81c8b] Took 1.68 seconds to deallocate network for instance. [ 931.162017] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.162598] env[62235]: DEBUG nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.167757] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.961s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.167757] env[62235]: INFO nova.compute.claims [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.335182] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.509200] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.509439] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.674804] env[62235]: DEBUG nova.compute.utils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.677067] env[62235]: DEBUG nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.677145] env[62235]: DEBUG nova.network.neutron [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 931.717638] env[62235]: DEBUG nova.policy [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '738ee55c831c48f894efbcef71bc20a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c943a445f67f4021bad11723e0836537', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.014029] env[62235]: DEBUG nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.022257] env[62235]: DEBUG nova.network.neutron [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Successfully created port: c536912e-8bb6-44c3-9ab4-fdc21a8c1556 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.102797] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0d3fdc9-1b9b-432b-a1a8-6ace10473973 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d98ab48f-c325-473a-8ee8-57ecd3b81c8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.762s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.181070] env[62235]: DEBUG nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.235227] env[62235]: DEBUG nova.compute.manager [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Stashing vm_state: active {{(pid=62235) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 932.380334] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17da40ad-03dd-4b88-87aa-6f66906f8682 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.388158] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c6dec3-d644-45ea-90b5-a6728ebace4a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.418962] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66863ba-2cac-4dec-95e8-d364c438be3f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.426314] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e04d73e-4a47-4df6-a320-0ad9497392d7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.439486] env[62235]: DEBUG nova.compute.provider_tree [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.530486] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.752890] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.789192] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "6f803e24-2b2b-4a15-8172-29acde2154f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.789432] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.942996] env[62235]: DEBUG nova.scheduler.client.report [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.189461] env[62235]: DEBUG nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.228802] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.229080] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.229250] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.229448] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.229598] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.229751] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.229964] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.230146] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.230373] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.230513] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.230695] env[62235]: DEBUG nova.virt.hardware [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.231597] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0bb834-9bd4-41ea-86c2-10df7b5f66b4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.239808] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ac29ed-47b6-4dc3-9a6a-e73d5db1c897 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.292400] env[62235]: DEBUG nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 933.448187] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.448885] env[62235]: DEBUG nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.452249] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.940s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.452732] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.452732] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 933.452914] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.936s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.454267] env[62235]: INFO nova.compute.claims [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.458648] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eebe976-10ca-4b37-951a-2aedd49755e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.467028] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25c7371-4e2b-4048-85d3-60d821e9d149 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.481011] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c0436d-6fb0-4d28-b02e-4d0efea0421b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.487451] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df5fe4a-f3f8-49ad-bc91-6a3b7aa737d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.519666] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180381MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 933.519841] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.523753] env[62235]: DEBUG nova.compute.manager [req-26abb982-2354-42c8-8847-d782d9c9f999 req-d6031220-0ada-40e8-a1df-65f425bc4cac service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Received event network-vif-plugged-c536912e-8bb6-44c3-9ab4-fdc21a8c1556 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.523753] env[62235]: DEBUG oslo_concurrency.lockutils [req-26abb982-2354-42c8-8847-d782d9c9f999 req-d6031220-0ada-40e8-a1df-65f425bc4cac service nova] Acquiring lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.523753] env[62235]: DEBUG oslo_concurrency.lockutils [req-26abb982-2354-42c8-8847-d782d9c9f999 req-d6031220-0ada-40e8-a1df-65f425bc4cac service nova] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.523753] env[62235]: DEBUG oslo_concurrency.lockutils [req-26abb982-2354-42c8-8847-d782d9c9f999 req-d6031220-0ada-40e8-a1df-65f425bc4cac service nova] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.523753] env[62235]: DEBUG nova.compute.manager [req-26abb982-2354-42c8-8847-d782d9c9f999 req-d6031220-0ada-40e8-a1df-65f425bc4cac service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] No waiting events found dispatching network-vif-plugged-c536912e-8bb6-44c3-9ab4-fdc21a8c1556 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.523753] env[62235]: WARNING nova.compute.manager [req-26abb982-2354-42c8-8847-d782d9c9f999 req-d6031220-0ada-40e8-a1df-65f425bc4cac service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Received unexpected event network-vif-plugged-c536912e-8bb6-44c3-9ab4-fdc21a8c1556 for instance with vm_state building and task_state spawning. [ 933.592168] env[62235]: DEBUG nova.network.neutron [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Successfully updated port: c536912e-8bb6-44c3-9ab4-fdc21a8c1556 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.824565] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.962105] env[62235]: DEBUG nova.compute.utils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.963507] env[62235]: DEBUG nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.963683] env[62235]: DEBUG nova.network.neutron [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.006242] env[62235]: DEBUG nova.policy [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.094753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-0ac4202d-4e25-417b-9ab0-6429b73e8e98" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.094753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-0ac4202d-4e25-417b-9ab0-6429b73e8e98" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.094753] env[62235]: DEBUG nova.network.neutron [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.289974] env[62235]: DEBUG nova.network.neutron [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Successfully created port: 1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.468609] env[62235]: DEBUG nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.626210] env[62235]: DEBUG nova.network.neutron [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 934.649275] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1afb6e6-f103-4294-8f29-a97c63b73d04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.656837] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29744349-9c01-403e-bc92-4d95533a710b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.687230] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6aca3a-e338-4466-a618-9dd2ddf6815c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.694221] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd050ba1-d833-4ab5-8392-023c2b1cad3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.708286] env[62235]: DEBUG nova.compute.provider_tree [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.778796] env[62235]: DEBUG nova.network.neutron [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Updating instance_info_cache with network_info: [{"id": "c536912e-8bb6-44c3-9ab4-fdc21a8c1556", "address": "fa:16:3e:94:f2:e2", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc536912e-8b", "ovs_interfaceid": "c536912e-8bb6-44c3-9ab4-fdc21a8c1556", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.211075] env[62235]: DEBUG nova.scheduler.client.report [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.281756] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-0ac4202d-4e25-417b-9ab0-6429b73e8e98" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.282374] env[62235]: DEBUG nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Instance network_info: |[{"id": "c536912e-8bb6-44c3-9ab4-fdc21a8c1556", "address": "fa:16:3e:94:f2:e2", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc536912e-8b", "ovs_interfaceid": "c536912e-8bb6-44c3-9ab4-fdc21a8c1556", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 935.282537] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:f2:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37fb1918-d178-4e12-93e6-316381e78be4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c536912e-8bb6-44c3-9ab4-fdc21a8c1556', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.290102] env[62235]: DEBUG oslo.service.loopingcall [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.290325] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 935.290612] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-147fd863-e9a1-47ab-8988-e472b863195f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.312104] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.312104] env[62235]: value = "task-1271945" [ 935.312104] env[62235]: _type = "Task" [ 935.312104] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.320022] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271945, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.480699] env[62235]: DEBUG nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.509982] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.510290] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.510475] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.510695] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.510853] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.511112] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.511346] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.511541] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.511723] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.511894] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.512085] env[62235]: DEBUG nova.virt.hardware [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.512994] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd6373e-8e26-4704-b5f5-3daea6fdf338 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.521151] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b410122f-528a-413f-8931-183f18c0a6d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.606265] env[62235]: DEBUG nova.compute.manager [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Received event network-changed-c536912e-8bb6-44c3-9ab4-fdc21a8c1556 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.606479] env[62235]: DEBUG nova.compute.manager [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Refreshing instance network info cache due to event network-changed-c536912e-8bb6-44c3-9ab4-fdc21a8c1556. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.606716] env[62235]: DEBUG oslo_concurrency.lockutils [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] Acquiring lock "refresh_cache-0ac4202d-4e25-417b-9ab0-6429b73e8e98" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.606882] env[62235]: DEBUG oslo_concurrency.lockutils [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] Acquired lock "refresh_cache-0ac4202d-4e25-417b-9ab0-6429b73e8e98" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.607045] env[62235]: DEBUG nova.network.neutron [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Refreshing network info cache for port c536912e-8bb6-44c3-9ab4-fdc21a8c1556 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.716506] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.717061] env[62235]: DEBUG nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.720173] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.740s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.721814] env[62235]: INFO nova.compute.claims [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.823327] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271945, 'name': CreateVM_Task, 'duration_secs': 0.307989} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.823513] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 935.824202] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.824371] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.824685] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.825036] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfecad5b-9c17-424b-a662-c40018f7c186 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.829478] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 935.829478] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210b01f-565b-6e9e-57f0-e22774a6312a" [ 935.829478] env[62235]: _type = "Task" [ 935.829478] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.838799] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210b01f-565b-6e9e-57f0-e22774a6312a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.839599] env[62235]: DEBUG nova.network.neutron [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Successfully updated port: 1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.228672] env[62235]: DEBUG nova.compute.utils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.232031] env[62235]: DEBUG nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 936.232031] env[62235]: DEBUG nova.network.neutron [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.270894] env[62235]: DEBUG nova.policy [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31214d58a95c48f1b2a80911e3710b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfb228bc090c468d833754afe2bcb3ff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.319013] env[62235]: DEBUG nova.network.neutron [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Updated VIF entry in instance network info cache for port c536912e-8bb6-44c3-9ab4-fdc21a8c1556. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 936.319612] env[62235]: DEBUG nova.network.neutron [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Updating instance_info_cache with network_info: [{"id": "c536912e-8bb6-44c3-9ab4-fdc21a8c1556", "address": "fa:16:3e:94:f2:e2", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc536912e-8b", "ovs_interfaceid": "c536912e-8bb6-44c3-9ab4-fdc21a8c1556", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.341487] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210b01f-565b-6e9e-57f0-e22774a6312a, 'name': SearchDatastore_Task, 'duration_secs': 0.009144} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.342016] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.342176] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.342382] env[62235]: DEBUG nova.network.neutron [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.343427] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.343695] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.343935] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.344094] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.344287] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.344547] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3402fe13-6e67-4325-b046-59081d75f7c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.353873] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.354071] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.354854] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f89a29a-02be-4c6b-bf42-b7df283fa102 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.360322] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 936.360322] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2f710-5c9e-cdb8-ef65-f5193ff913c1" [ 936.360322] env[62235]: _type = "Task" [ 936.360322] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.367901] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2f710-5c9e-cdb8-ef65-f5193ff913c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.556429] env[62235]: DEBUG nova.network.neutron [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Successfully created port: f505e074-0353-4a28-abad-d058112f253a {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.735416] env[62235]: DEBUG nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 936.824141] env[62235]: DEBUG oslo_concurrency.lockutils [req-a195d1ac-8b70-4838-9199-d031503d107b req-83570f19-194d-4a95-9acf-4218029e28a1 service nova] Releasing lock "refresh_cache-0ac4202d-4e25-417b-9ab0-6429b73e8e98" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.870609] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2f710-5c9e-cdb8-ef65-f5193ff913c1, 'name': SearchDatastore_Task, 'duration_secs': 0.008172} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.873667] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-116bd0e2-3084-4f52-899e-0327a03e9e73 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.878086] env[62235]: DEBUG nova.network.neutron [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.881243] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 936.881243] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529b081e-8ded-d2b7-d75a-fa85cc6c7c96" [ 936.881243] env[62235]: _type = "Task" [ 936.881243] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.890488] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529b081e-8ded-d2b7-d75a-fa85cc6c7c96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.927786] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e33868-64f6-4bfc-87cf-da0dfbf92b80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.935030] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee4771f-6e09-47e7-bfd4-999a4053898b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.968723] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af361091-2a2c-4eea-9650-a0c8ca015a58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.976352] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd8a349-c24d-4bcb-9be0-d410f9a30ab1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.989831] env[62235]: DEBUG nova.compute.provider_tree [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.036696] env[62235]: DEBUG nova.network.neutron [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.392078] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529b081e-8ded-d2b7-d75a-fa85cc6c7c96, 'name': SearchDatastore_Task, 'duration_secs': 0.012044} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.392078] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.392475] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.392543] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-caad8297-4860-475a-a416-0fb9c95f28e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.400279] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 937.400279] env[62235]: value = "task-1271946" [ 937.400279] env[62235]: _type = "Task" [ 937.400279] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.408275] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271946, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.493360] env[62235]: DEBUG nova.scheduler.client.report [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.538886] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.539256] env[62235]: DEBUG nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Instance network_info: |[{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.539706] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:0d:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1bc70ab7-baf7-4b1d-b945-41463b6ea517', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.547458] env[62235]: DEBUG oslo.service.loopingcall [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.548119] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.548394] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3074300-010c-425d-b451-3b8e72fcaad2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.569252] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.569252] env[62235]: value = "task-1271947" [ 937.569252] env[62235]: _type = "Task" [ 937.569252] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.577445] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271947, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.635458] env[62235]: DEBUG nova.compute.manager [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-vif-plugged-1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.636097] env[62235]: DEBUG oslo_concurrency.lockutils [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] Acquiring lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.636097] env[62235]: DEBUG oslo_concurrency.lockutils [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.636097] env[62235]: DEBUG oslo_concurrency.lockutils [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.636304] env[62235]: DEBUG nova.compute.manager [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] No waiting events found dispatching network-vif-plugged-1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.636453] env[62235]: WARNING nova.compute.manager [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received unexpected event network-vif-plugged-1bc70ab7-baf7-4b1d-b945-41463b6ea517 for instance with vm_state building and task_state spawning. [ 937.636702] env[62235]: DEBUG nova.compute.manager [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.636874] env[62235]: DEBUG nova.compute.manager [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing instance network info cache due to event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.637281] env[62235]: DEBUG oslo_concurrency.lockutils [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.637464] env[62235]: DEBUG oslo_concurrency.lockutils [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.637641] env[62235]: DEBUG nova.network.neutron [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.750766] env[62235]: DEBUG nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 937.783516] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.783837] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.784028] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.784261] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.784440] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.784616] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.784860] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.785140] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.785380] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.785593] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.785817] env[62235]: DEBUG nova.virt.hardware [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.787147] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da34f578-9c1c-4a92-a880-7456dfa0a708 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.796567] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48384edc-a0f4-45b7-8ef4-99cfbfae967b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.909442] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271946, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436334} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.909727] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 937.909945] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 937.910474] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6203c22-1438-4f3f-bcad-66720d2d1a55 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.920249] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 937.920249] env[62235]: value = "task-1271948" [ 937.920249] env[62235]: _type = "Task" [ 937.920249] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.927789] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271948, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.988577] env[62235]: DEBUG nova.compute.manager [req-cf8e9dc2-418f-49ed-84e9-92f93d153838 req-eaaa470c-8e87-42dd-bef7-87267425b64c service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Received event network-vif-plugged-f505e074-0353-4a28-abad-d058112f253a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.988814] env[62235]: DEBUG oslo_concurrency.lockutils [req-cf8e9dc2-418f-49ed-84e9-92f93d153838 req-eaaa470c-8e87-42dd-bef7-87267425b64c service nova] Acquiring lock "eccde807-664e-4488-bec8-2d6919c2fc91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.989053] env[62235]: DEBUG oslo_concurrency.lockutils [req-cf8e9dc2-418f-49ed-84e9-92f93d153838 req-eaaa470c-8e87-42dd-bef7-87267425b64c service nova] Lock "eccde807-664e-4488-bec8-2d6919c2fc91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.989247] env[62235]: DEBUG oslo_concurrency.lockutils [req-cf8e9dc2-418f-49ed-84e9-92f93d153838 req-eaaa470c-8e87-42dd-bef7-87267425b64c service nova] Lock "eccde807-664e-4488-bec8-2d6919c2fc91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.989420] env[62235]: DEBUG nova.compute.manager [req-cf8e9dc2-418f-49ed-84e9-92f93d153838 req-eaaa470c-8e87-42dd-bef7-87267425b64c service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] No waiting events found dispatching network-vif-plugged-f505e074-0353-4a28-abad-d058112f253a {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.989585] env[62235]: WARNING nova.compute.manager [req-cf8e9dc2-418f-49ed-84e9-92f93d153838 req-eaaa470c-8e87-42dd-bef7-87267425b64c service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Received unexpected event network-vif-plugged-f505e074-0353-4a28-abad-d058112f253a for instance with vm_state building and task_state spawning. [ 937.999215] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.999672] env[62235]: DEBUG nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.002580] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.668s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.003706] env[62235]: INFO nova.compute.claims [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.079342] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271947, 'name': CreateVM_Task, 'duration_secs': 0.376921} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.080020] env[62235]: DEBUG nova.network.neutron [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Successfully updated port: f505e074-0353-4a28-abad-d058112f253a {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.081037] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.081931] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.082058] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.082340] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.082970] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ede701a-6c1e-487f-ad77-bd8955e3e737 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.086947] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 938.086947] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5261bfa9-afc0-2329-262a-c3b9ff9333a2" [ 938.086947] env[62235]: _type = "Task" [ 938.086947] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.095489] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5261bfa9-afc0-2329-262a-c3b9ff9333a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.324429] env[62235]: DEBUG nova.network.neutron [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updated VIF entry in instance network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.324833] env[62235]: DEBUG nova.network.neutron [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.428727] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271948, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061013} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.429035] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 938.429797] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d7c213-cccf-420d-81c6-9f4809e3a4ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.451461] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.451732] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0fb02e5-3a98-4725-ac66-63bc0a2043b0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.474271] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 938.474271] env[62235]: value = "task-1271949" [ 938.474271] env[62235]: _type = "Task" [ 938.474271] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.481747] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271949, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.508224] env[62235]: DEBUG nova.compute.utils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.513088] env[62235]: DEBUG nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.513276] env[62235]: DEBUG nova.network.neutron [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 938.550099] env[62235]: DEBUG nova.policy [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31214d58a95c48f1b2a80911e3710b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfb228bc090c468d833754afe2bcb3ff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.582240] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.582321] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.582501] env[62235]: DEBUG nova.network.neutron [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 938.599597] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5261bfa9-afc0-2329-262a-c3b9ff9333a2, 'name': SearchDatastore_Task, 'duration_secs': 0.009194} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.599942] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.600190] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.600424] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.600598] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.600793] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.601077] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-959e9061-7870-4a53-b1ae-b7f3a07fe558 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.609435] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.609632] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.610335] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3ee8903-9ba0-4d41-9fb7-4587abc682d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.615400] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 938.615400] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521d78ef-1c11-346f-969a-f16b0957161e" [ 938.615400] env[62235]: _type = "Task" [ 938.615400] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.623081] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521d78ef-1c11-346f-969a-f16b0957161e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.827296] env[62235]: DEBUG oslo_concurrency.lockutils [req-db15e318-ec28-44aa-86ae-0fcc6a147423 req-faf3274b-f162-49f1-ba73-6289fe96a124 service nova] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.961569] env[62235]: DEBUG nova.network.neutron [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Successfully created port: 21adaaf5-f619-4e50-9341-0cb69b5c91c6 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.985469] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271949, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.011318] env[62235]: DEBUG nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.116598] env[62235]: DEBUG nova.network.neutron [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 939.129766] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521d78ef-1c11-346f-969a-f16b0957161e, 'name': SearchDatastore_Task, 'duration_secs': 0.008077} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.130333] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5b68cb0-7957-42c2-b5c4-fa05b8bcda04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.138250] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 939.138250] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52889104-824c-bc13-9bb5-47435e9c2d48" [ 939.138250] env[62235]: _type = "Task" [ 939.138250] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.146781] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52889104-824c-bc13-9bb5-47435e9c2d48, 'name': SearchDatastore_Task, 'duration_secs': 0.008438} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.148995] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.149282] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a3336b20-5737-4e5a-94a1-19fec140aa60/a3336b20-5737-4e5a-94a1-19fec140aa60.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 939.150091] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55ca0393-1e93-4067-bc6a-41089a169760 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.156152] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 939.156152] env[62235]: value = "task-1271950" [ 939.156152] env[62235]: _type = "Task" [ 939.156152] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.167474] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271950, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.216649] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53684320-efec-4bf5-a203-3e038c46d534 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.224451] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d55e556-1622-4d24-9e8d-f8ae9002a2fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.257584] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97302f84-12ad-462f-b964-0a44755dedd3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.264729] env[62235]: DEBUG nova.network.neutron [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Updating instance_info_cache with network_info: [{"id": "f505e074-0353-4a28-abad-d058112f253a", "address": "fa:16:3e:20:74:13", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf505e074-03", "ovs_interfaceid": "f505e074-0353-4a28-abad-d058112f253a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.266886] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775606de-1c78-4562-b038-606ec64ce0b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.284148] env[62235]: DEBUG nova.compute.provider_tree [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.487957] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271949, 'name': ReconfigVM_Task, 'duration_secs': 0.604875} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.488394] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.489200] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-227c031e-4b6c-412a-9af9-01e94fedbd82 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.496519] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 939.496519] env[62235]: value = "task-1271951" [ 939.496519] env[62235]: _type = "Task" [ 939.496519] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.506321] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271951, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.665787] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271950, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448284} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.666026] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a3336b20-5737-4e5a-94a1-19fec140aa60/a3336b20-5737-4e5a-94a1-19fec140aa60.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 939.666271] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.666531] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa13cd38-092f-4f3b-9490-ddcc362788fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.673516] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 939.673516] env[62235]: value = "task-1271952" [ 939.673516] env[62235]: _type = "Task" [ 939.673516] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.681747] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.775296] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.775785] env[62235]: DEBUG nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Instance network_info: |[{"id": "f505e074-0353-4a28-abad-d058112f253a", "address": "fa:16:3e:20:74:13", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf505e074-03", "ovs_interfaceid": "f505e074-0353-4a28-abad-d058112f253a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 939.776244] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:74:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f505e074-0353-4a28-abad-d058112f253a', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.783989] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Creating folder: Project (bfb228bc090c468d833754afe2bcb3ff). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 939.784386] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-572bb86b-3347-46e9-b522-be9e4c6ee601 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.786872] env[62235]: DEBUG nova.scheduler.client.report [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.799727] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Created folder: Project (bfb228bc090c468d833754afe2bcb3ff) in parent group-v273362. [ 939.799987] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Creating folder: Instances. Parent ref: group-v273491. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 939.801080] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-166f901c-b9e3-49ea-a5ee-6417970e97fb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.811246] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Created folder: Instances in parent group-v273491. [ 939.811541] env[62235]: DEBUG oslo.service.loopingcall [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.811761] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 939.811988] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63df616c-5591-45ee-a695-2d7224cf18b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.832582] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.832582] env[62235]: value = "task-1271955" [ 939.832582] env[62235]: _type = "Task" [ 939.832582] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.841412] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271955, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.006845] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271951, 'name': Rename_Task, 'duration_secs': 0.166774} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.007889] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.007889] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63f74630-eeb9-4b18-a55f-d22f5207d299 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.014305] env[62235]: DEBUG nova.compute.manager [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Received event network-changed-f505e074-0353-4a28-abad-d058112f253a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.014501] env[62235]: DEBUG nova.compute.manager [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Refreshing instance network info cache due to event network-changed-f505e074-0353-4a28-abad-d058112f253a. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.014721] env[62235]: DEBUG oslo_concurrency.lockutils [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] Acquiring lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.014863] env[62235]: DEBUG oslo_concurrency.lockutils [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] Acquired lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.015149] env[62235]: DEBUG nova.network.neutron [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Refreshing network info cache for port f505e074-0353-4a28-abad-d058112f253a {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.017594] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 940.017594] env[62235]: value = "task-1271956" [ 940.017594] env[62235]: _type = "Task" [ 940.017594] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.027156] env[62235]: DEBUG nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.029381] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271956, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.052440] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.052710] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.052876] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.053078] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.053236] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.053390] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.053601] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.053804] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.053995] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.054189] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.054367] env[62235]: DEBUG nova.virt.hardware [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.055532] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6981fc43-e8c0-4e88-9ada-71258179f18d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.063765] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52db53fd-4aba-4276-b76f-cf992cfc8587 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.182872] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068649} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.183173] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.183945] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d4894d-8aca-46d7-a182-c51254c91617 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.205810] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] a3336b20-5737-4e5a-94a1-19fec140aa60/a3336b20-5737-4e5a-94a1-19fec140aa60.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.206209] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-495dcea0-b138-4f00-a474-ceb2918509d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.226941] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 940.226941] env[62235]: value = "task-1271957" [ 940.226941] env[62235]: _type = "Task" [ 940.226941] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.236788] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.293029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.293029] env[62235]: DEBUG nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 940.295408] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.765s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.297146] env[62235]: INFO nova.compute.claims [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.343444] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271955, 'name': CreateVM_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.492388] env[62235]: DEBUG nova.network.neutron [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Successfully updated port: 21adaaf5-f619-4e50-9341-0cb69b5c91c6 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.530438] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271956, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.732742] env[62235]: DEBUG nova.network.neutron [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Updated VIF entry in instance network info cache for port f505e074-0353-4a28-abad-d058112f253a. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 940.733137] env[62235]: DEBUG nova.network.neutron [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Updating instance_info_cache with network_info: [{"id": "f505e074-0353-4a28-abad-d058112f253a", "address": "fa:16:3e:20:74:13", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf505e074-03", "ovs_interfaceid": "f505e074-0353-4a28-abad-d058112f253a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.737799] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.802704] env[62235]: DEBUG nova.compute.utils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.808237] env[62235]: DEBUG nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.808555] env[62235]: DEBUG nova.network.neutron [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 940.842796] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271955, 'name': CreateVM_Task, 'duration_secs': 0.814666} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.842971] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.843673] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.843845] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.844195] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.844460] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4348bcfe-cb34-4aa7-b0f4-d67de0f0d507 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.849094] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 940.849094] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52728e7e-c547-c201-4ef2-21cb447196c8" [ 940.849094] env[62235]: _type = "Task" [ 940.849094] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.850432] env[62235]: DEBUG nova.policy [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31214d58a95c48f1b2a80911e3710b36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfb228bc090c468d833754afe2bcb3ff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.860185] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52728e7e-c547-c201-4ef2-21cb447196c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.997732] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "refresh_cache-85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.997732] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "refresh_cache-85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.997732] env[62235]: DEBUG nova.network.neutron [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.032950] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271956, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.135910] env[62235]: DEBUG nova.network.neutron [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Successfully created port: e05bb07a-f9f3-48a9-847a-0b12ba578ea8 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.239707] env[62235]: DEBUG oslo_concurrency.lockutils [req-786b5974-234c-42e5-afe5-cceadb020cf8 req-03d21ebe-c9fe-4e69-ba0b-96826b6cb657 service nova] Releasing lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.240137] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271957, 'name': ReconfigVM_Task, 'duration_secs': 0.824803} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.240474] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Reconfigured VM instance instance-00000056 to attach disk [datastore2] a3336b20-5737-4e5a-94a1-19fec140aa60/a3336b20-5737-4e5a-94a1-19fec140aa60.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.241031] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1112b189-90dd-4809-bebe-ee4b077c484d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.247972] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 941.247972] env[62235]: value = "task-1271958" [ 941.247972] env[62235]: _type = "Task" [ 941.247972] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.255920] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271958, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.309339] env[62235]: DEBUG nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 941.362388] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52728e7e-c547-c201-4ef2-21cb447196c8, 'name': SearchDatastore_Task, 'duration_secs': 0.015957} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.365033] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.365289] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.365528] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.365681] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.365869] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.366329] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0bb38b1-20c9-49dc-8873-17c675527e64 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.375556] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.375960] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.379073] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68d305a3-0a6c-49ff-afc3-a926e2a0f540 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.384554] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 941.384554] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fec1b2-d232-054a-29f2-c72798130185" [ 941.384554] env[62235]: _type = "Task" [ 941.384554] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.392090] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fec1b2-d232-054a-29f2-c72798130185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.491617] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e353126f-9e58-482a-922a-1a6a6c43cac1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.499159] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23894696-9ccc-455a-ae2b-da590f5f11dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.532977] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba126c7a-d1be-4895-b168-5c43f047222c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.535909] env[62235]: DEBUG nova.network.neutron [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.542344] env[62235]: DEBUG oslo_vmware.api [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271956, 'name': PowerOnVM_Task, 'duration_secs': 1.052121} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.544372] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.544600] env[62235]: INFO nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Took 8.35 seconds to spawn the instance on the hypervisor. [ 941.544793] env[62235]: DEBUG nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.545589] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728cc78d-185a-4b93-bc3a-c350bcaa97d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.548830] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b88d96-7f7c-4d07-a02c-4183896c3bfc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.565520] env[62235]: DEBUG nova.compute.provider_tree [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.680822] env[62235]: DEBUG nova.network.neutron [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Updating instance_info_cache with network_info: [{"id": "21adaaf5-f619-4e50-9341-0cb69b5c91c6", "address": "fa:16:3e:0d:21:a2", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21adaaf5-f6", "ovs_interfaceid": "21adaaf5-f619-4e50-9341-0cb69b5c91c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.757745] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271958, 'name': Rename_Task, 'duration_secs': 0.128541} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.758057] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.758321] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb7c2ad3-0fad-4838-9da6-916c4849397d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.765590] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 941.765590] env[62235]: value = "task-1271959" [ 941.765590] env[62235]: _type = "Task" [ 941.765590] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.773268] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271959, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.895897] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fec1b2-d232-054a-29f2-c72798130185, 'name': SearchDatastore_Task, 'duration_secs': 0.009509} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.896080] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b208fa77-4bf5-482c-bb59-5c7d7998dfec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.900818] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 941.900818] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521557bb-8ef5-ff68-78d6-9887bd8b7006" [ 941.900818] env[62235]: _type = "Task" [ 941.900818] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.907908] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521557bb-8ef5-ff68-78d6-9887bd8b7006, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.042927] env[62235]: DEBUG nova.compute.manager [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Received event network-vif-plugged-21adaaf5-f619-4e50-9341-0cb69b5c91c6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.043274] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] Acquiring lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.043396] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.043570] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.043748] env[62235]: DEBUG nova.compute.manager [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] No waiting events found dispatching network-vif-plugged-21adaaf5-f619-4e50-9341-0cb69b5c91c6 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.043919] env[62235]: WARNING nova.compute.manager [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Received unexpected event network-vif-plugged-21adaaf5-f619-4e50-9341-0cb69b5c91c6 for instance with vm_state building and task_state spawning. [ 942.044098] env[62235]: DEBUG nova.compute.manager [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Received event network-changed-21adaaf5-f619-4e50-9341-0cb69b5c91c6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.044256] env[62235]: DEBUG nova.compute.manager [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Refreshing instance network info cache due to event network-changed-21adaaf5-f619-4e50-9341-0cb69b5c91c6. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.044456] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] Acquiring lock "refresh_cache-85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.075022] env[62235]: DEBUG nova.scheduler.client.report [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.083429] env[62235]: INFO nova.compute.manager [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Took 14.18 seconds to build instance. [ 942.184127] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "refresh_cache-85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.184620] env[62235]: DEBUG nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Instance network_info: |[{"id": "21adaaf5-f619-4e50-9341-0cb69b5c91c6", "address": "fa:16:3e:0d:21:a2", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21adaaf5-f6", "ovs_interfaceid": "21adaaf5-f619-4e50-9341-0cb69b5c91c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.185120] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] Acquired lock "refresh_cache-85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.185457] env[62235]: DEBUG nova.network.neutron [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Refreshing network info cache for port 21adaaf5-f619-4e50-9341-0cb69b5c91c6 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.188137] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:21:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21adaaf5-f619-4e50-9341-0cb69b5c91c6', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.194691] env[62235]: DEBUG oslo.service.loopingcall [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.195819] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.196134] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-840ada30-4bbd-477f-be2a-f8944b97c925 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.221974] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.221974] env[62235]: value = "task-1271960" [ 942.221974] env[62235]: _type = "Task" [ 942.221974] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.232268] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271960, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.275948] env[62235]: DEBUG oslo_vmware.api [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1271959, 'name': PowerOnVM_Task, 'duration_secs': 0.468324} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.275948] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 942.275948] env[62235]: INFO nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Took 6.79 seconds to spawn the instance on the hypervisor. [ 942.275948] env[62235]: DEBUG nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.276563] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f18efa-1b89-4a98-b606-3b641799bec8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.322821] env[62235]: DEBUG nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 942.356289] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.356642] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.356854] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.357094] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.357301] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.357956] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.358398] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.358542] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.358793] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.359056] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.359433] env[62235]: DEBUG nova.virt.hardware [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.360822] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b413a7e3-6686-4277-b824-27f8345d8f89 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.369626] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239ec0b2-12f6-4f15-aa4a-6eaa147c73bd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.411413] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521557bb-8ef5-ff68-78d6-9887bd8b7006, 'name': SearchDatastore_Task, 'duration_secs': 0.008611} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.411778] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.412097] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] eccde807-664e-4488-bec8-2d6919c2fc91/eccde807-664e-4488-bec8-2d6919c2fc91.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.412396] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9301e88-84e4-4b87-8975-e5f2e936bed1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.419443] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 942.419443] env[62235]: value = "task-1271961" [ 942.419443] env[62235]: _type = "Task" [ 942.419443] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.427244] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271961, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.581617] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.582256] env[62235]: DEBUG nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.584849] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 9.832s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.586422] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c1cffd22-3c12-40a1-87dc-947864fafdf3 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.688s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.733079] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271960, 'name': CreateVM_Task, 'duration_secs': 0.358125} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.735738] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.740544] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.740544] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.740544] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.740544] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27d08eaf-27be-4aaf-b560-aab3d3800e29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.745358] env[62235]: DEBUG nova.network.neutron [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Successfully updated port: e05bb07a-f9f3-48a9-847a-0b12ba578ea8 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.747501] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 942.747501] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52355412-47e0-2a89-22ea-a5e35e94ba80" [ 942.747501] env[62235]: _type = "Task" [ 942.747501] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.759275] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52355412-47e0-2a89-22ea-a5e35e94ba80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.796557] env[62235]: INFO nova.compute.manager [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Took 14.61 seconds to build instance. [ 942.929939] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271961, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502179} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.930263] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] eccde807-664e-4488-bec8-2d6919c2fc91/eccde807-664e-4488-bec8-2d6919c2fc91.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.930516] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.930808] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a5513b2-befc-48ed-a4a9-82d66347acea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.937123] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 942.937123] env[62235]: value = "task-1271962" [ 942.937123] env[62235]: _type = "Task" [ 942.937123] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.946050] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271962, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.947421] env[62235]: DEBUG nova.network.neutron [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Updated VIF entry in instance network info cache for port 21adaaf5-f619-4e50-9341-0cb69b5c91c6. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.948129] env[62235]: DEBUG nova.network.neutron [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Updating instance_info_cache with network_info: [{"id": "21adaaf5-f619-4e50-9341-0cb69b5c91c6", "address": "fa:16:3e:0d:21:a2", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21adaaf5-f6", "ovs_interfaceid": "21adaaf5-f619-4e50-9341-0cb69b5c91c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.090152] env[62235]: INFO nova.compute.claims [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.094738] env[62235]: DEBUG nova.compute.utils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.096373] env[62235]: DEBUG nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.096534] env[62235]: DEBUG nova.network.neutron [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 943.126490] env[62235]: INFO nova.compute.manager [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Rebuilding instance [ 943.136031] env[62235]: DEBUG nova.policy [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11fe623e504d4729a149cf666aed4c67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b690e5ef3dac4a07879b4b7fe2249e89', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 943.174325] env[62235]: DEBUG nova.compute.manager [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.175241] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed27dabf-d8c0-42b2-85e4-8ef2ae944331 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.249725] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "refresh_cache-79dafae1-e389-4e04-af20-577b8aac06a6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.249725] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "refresh_cache-79dafae1-e389-4e04-af20-577b8aac06a6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.249725] env[62235]: DEBUG nova.network.neutron [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.266161] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52355412-47e0-2a89-22ea-a5e35e94ba80, 'name': SearchDatastore_Task, 'duration_secs': 0.024151} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.266481] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.266713] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.266985] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.267159] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.267348] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.267608] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-423220e5-4188-4339-bc1d-ab66b07fe45c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.275618] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.275799] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.276495] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e955df35-a7de-420e-a4b1-e9a9c592b257 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.281553] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 943.281553] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523f4622-146a-c2bd-c79c-48b3c28bfb19" [ 943.281553] env[62235]: _type = "Task" [ 943.281553] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.289013] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523f4622-146a-c2bd-c79c-48b3c28bfb19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.298813] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6735a76-8e6d-4dc5-a692-2494c0bf4a48 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.123s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.383642] env[62235]: DEBUG nova.network.neutron [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Successfully created port: 7d53cc57-246b-4fed-a55d-514cd12548d8 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.446430] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271962, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062302} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.446708] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.447537] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd218da-5c8c-4c3f-b0f9-d97d93882fe7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.451612] env[62235]: DEBUG oslo_concurrency.lockutils [req-7e05e243-7130-48c3-9b61-2d2dda0926a0 req-c8fbb1ad-6926-4860-926e-44f4ba16a9e9 service nova] Releasing lock "refresh_cache-85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.469508] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] eccde807-664e-4488-bec8-2d6919c2fc91/eccde807-664e-4488-bec8-2d6919c2fc91.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.469798] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-853619f7-4518-4783-abab-53e802245e5a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.491996] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 943.491996] env[62235]: value = "task-1271963" [ 943.491996] env[62235]: _type = "Task" [ 943.491996] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.502321] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271963, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.599277] env[62235]: INFO nova.compute.resource_tracker [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating resource usage from migration b1367deb-b42b-418c-b261-69119a3fb577 [ 943.601739] env[62235]: DEBUG nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.689129] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.689833] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a722a99-535a-4380-aa32-9f6089b05d22 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.696866] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 943.696866] env[62235]: value = "task-1271964" [ 943.696866] env[62235]: _type = "Task" [ 943.696866] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.706044] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.792480] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523f4622-146a-c2bd-c79c-48b3c28bfb19, 'name': SearchDatastore_Task, 'duration_secs': 0.012011} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.793281] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9c6c47c-25b6-4d95-9108-f617a86717ce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.800579] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 943.800579] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f34b2f-e20b-1f34-e820-50eeaefbf106" [ 943.800579] env[62235]: _type = "Task" [ 943.800579] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.810119] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f34b2f-e20b-1f34-e820-50eeaefbf106, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.811741] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2170f3-cddc-400b-9f75-c7779827cdfd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.818431] env[62235]: DEBUG nova.network.neutron [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.826887] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2836865-0207-4a44-98d8-f63c4e683ac0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.870191] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83e21b0-e1bf-4996-8261-d79a8ed74af8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.878992] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0739117f-599b-4406-a65e-fd8b576da2d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.893935] env[62235]: DEBUG nova.compute.provider_tree [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.001331] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271963, 'name': ReconfigVM_Task, 'duration_secs': 0.475225} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.001575] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Reconfigured VM instance instance-00000057 to attach disk [datastore2] eccde807-664e-4488-bec8-2d6919c2fc91/eccde807-664e-4488-bec8-2d6919c2fc91.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.002217] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-401420df-3c43-49c2-a623-09b40878baaa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.007630] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 944.007630] env[62235]: value = "task-1271965" [ 944.007630] env[62235]: _type = "Task" [ 944.007630] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.015105] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271965, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.075067] env[62235]: DEBUG nova.compute.manager [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Received event network-vif-plugged-e05bb07a-f9f3-48a9-847a-0b12ba578ea8 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.075384] env[62235]: DEBUG oslo_concurrency.lockutils [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] Acquiring lock "79dafae1-e389-4e04-af20-577b8aac06a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.075552] env[62235]: DEBUG oslo_concurrency.lockutils [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] Lock "79dafae1-e389-4e04-af20-577b8aac06a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.075729] env[62235]: DEBUG oslo_concurrency.lockutils [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] Lock "79dafae1-e389-4e04-af20-577b8aac06a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.075918] env[62235]: DEBUG nova.compute.manager [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] No waiting events found dispatching network-vif-plugged-e05bb07a-f9f3-48a9-847a-0b12ba578ea8 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 944.076119] env[62235]: WARNING nova.compute.manager [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Received unexpected event network-vif-plugged-e05bb07a-f9f3-48a9-847a-0b12ba578ea8 for instance with vm_state building and task_state spawning. [ 944.076302] env[62235]: DEBUG nova.compute.manager [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Received event network-changed-e05bb07a-f9f3-48a9-847a-0b12ba578ea8 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.076743] env[62235]: DEBUG nova.compute.manager [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Refreshing instance network info cache due to event network-changed-e05bb07a-f9f3-48a9-847a-0b12ba578ea8. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.076992] env[62235]: DEBUG oslo_concurrency.lockutils [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] Acquiring lock "refresh_cache-79dafae1-e389-4e04-af20-577b8aac06a6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.114058] env[62235]: DEBUG nova.network.neutron [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Updating instance_info_cache with network_info: [{"id": "e05bb07a-f9f3-48a9-847a-0b12ba578ea8", "address": "fa:16:3e:9b:a7:be", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape05bb07a-f9", "ovs_interfaceid": "e05bb07a-f9f3-48a9-847a-0b12ba578ea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.206535] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271964, 'name': PowerOffVM_Task, 'duration_secs': 0.260178} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.206817] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.207076] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.207827] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b8408e-a6f0-4fc3-be82-9c4bf1bcc079 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.215234] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 944.215460] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b4d5808-97fe-453c-9e28-be56d0e6224f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.285544] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 944.285808] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 944.286107] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleting the datastore file [datastore2] 0ac4202d-4e25-417b-9ab0-6429b73e8e98 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.286251] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32f3b694-6f27-4d17-8329-e55363768bdb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.293493] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 944.293493] env[62235]: value = "task-1271967" [ 944.293493] env[62235]: _type = "Task" [ 944.293493] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.303660] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.311374] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f34b2f-e20b-1f34-e820-50eeaefbf106, 'name': SearchDatastore_Task, 'duration_secs': 0.009191} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.311623] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.311883] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf/85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.312147] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a887f43a-67b9-4b14-8170-d6ce4bcb7403 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.317560] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 944.317560] env[62235]: value = "task-1271968" [ 944.317560] env[62235]: _type = "Task" [ 944.317560] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.324845] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271968, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.397983] env[62235]: DEBUG nova.scheduler.client.report [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.519316] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271965, 'name': Rename_Task, 'duration_secs': 0.133755} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.519643] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.519909] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17d1a1b4-da80-43b9-91ac-f63b1f298771 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.527264] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 944.527264] env[62235]: value = "task-1271969" [ 944.527264] env[62235]: _type = "Task" [ 944.527264] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.535841] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.615624] env[62235]: DEBUG nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.618446] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "refresh_cache-79dafae1-e389-4e04-af20-577b8aac06a6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.618975] env[62235]: DEBUG nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Instance network_info: |[{"id": "e05bb07a-f9f3-48a9-847a-0b12ba578ea8", "address": "fa:16:3e:9b:a7:be", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape05bb07a-f9", "ovs_interfaceid": "e05bb07a-f9f3-48a9-847a-0b12ba578ea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 944.619154] env[62235]: DEBUG oslo_concurrency.lockutils [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] Acquired lock "refresh_cache-79dafae1-e389-4e04-af20-577b8aac06a6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.619372] env[62235]: DEBUG nova.network.neutron [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Refreshing network info cache for port e05bb07a-f9f3-48a9-847a-0b12ba578ea8 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.620811] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:a7:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e05bb07a-f9f3-48a9-847a-0b12ba578ea8', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.630340] env[62235]: DEBUG oslo.service.loopingcall [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.631895] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 944.632320] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d716e7e8-e675-42e9-bac9-e3b5d1af72c6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.658423] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.658756] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.658906] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.659181] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.659399] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.659597] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.659851] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.660065] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.660305] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.661063] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.661063] env[62235]: DEBUG nova.virt.hardware [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.662206] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77934ff-49c8-46a4-8cc5-e0e8c16bdb4c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.667032] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.667032] env[62235]: value = "task-1271970" [ 944.667032] env[62235]: _type = "Task" [ 944.667032] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.675227] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d12945-2aba-475f-a39f-9a6f456e66c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.683321] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271970, 'name': CreateVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.805069] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180496} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.805394] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.805597] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 944.805807] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 944.827093] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271968, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488475} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.827398] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf/85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.828512] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.828512] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-259fbb25-32a6-4171-af05-8f44055eee51 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.834513] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 944.834513] env[62235]: value = "task-1271971" [ 944.834513] env[62235]: _type = "Task" [ 944.834513] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.842930] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271971, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.903680] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.319s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.903939] env[62235]: INFO nova.compute.manager [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Migrating [ 944.912521] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.393s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.918355] env[62235]: DEBUG nova.compute.manager [req-6be8b489-2b00-4fd2-995b-80119b6b7aac req-5dd426a4-82c3-4f8d-b39b-4984f106264d service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Received event network-vif-plugged-7d53cc57-246b-4fed-a55d-514cd12548d8 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.918355] env[62235]: DEBUG oslo_concurrency.lockutils [req-6be8b489-2b00-4fd2-995b-80119b6b7aac req-5dd426a4-82c3-4f8d-b39b-4984f106264d service nova] Acquiring lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.918355] env[62235]: DEBUG oslo_concurrency.lockutils [req-6be8b489-2b00-4fd2-995b-80119b6b7aac req-5dd426a4-82c3-4f8d-b39b-4984f106264d service nova] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.918355] env[62235]: DEBUG oslo_concurrency.lockutils [req-6be8b489-2b00-4fd2-995b-80119b6b7aac req-5dd426a4-82c3-4f8d-b39b-4984f106264d service nova] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.918355] env[62235]: DEBUG nova.compute.manager [req-6be8b489-2b00-4fd2-995b-80119b6b7aac req-5dd426a4-82c3-4f8d-b39b-4984f106264d service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] No waiting events found dispatching network-vif-plugged-7d53cc57-246b-4fed-a55d-514cd12548d8 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 944.918355] env[62235]: WARNING nova.compute.manager [req-6be8b489-2b00-4fd2-995b-80119b6b7aac req-5dd426a4-82c3-4f8d-b39b-4984f106264d service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Received unexpected event network-vif-plugged-7d53cc57-246b-4fed-a55d-514cd12548d8 for instance with vm_state building and task_state spawning. [ 945.037700] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271969, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.044807] env[62235]: DEBUG nova.network.neutron [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Successfully updated port: 7d53cc57-246b-4fed-a55d-514cd12548d8 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.177916] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271970, 'name': CreateVM_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.344664] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271971, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087936} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.345025] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.345905] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66cc596b-ef28-400e-a441-168a480e4d25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.370974] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf/85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.373197] env[62235]: DEBUG nova.network.neutron [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Updated VIF entry in instance network info cache for port e05bb07a-f9f3-48a9-847a-0b12ba578ea8. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.373197] env[62235]: DEBUG nova.network.neutron [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Updating instance_info_cache with network_info: [{"id": "e05bb07a-f9f3-48a9-847a-0b12ba578ea8", "address": "fa:16:3e:9b:a7:be", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape05bb07a-f9", "ovs_interfaceid": "e05bb07a-f9f3-48a9-847a-0b12ba578ea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.374799] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d591709-9709-4043-aa9b-7203f079e335 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.400020] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 945.400020] env[62235]: value = "task-1271972" [ 945.400020] env[62235]: _type = "Task" [ 945.400020] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.407019] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271972, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.430374] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.430561] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.430845] env[62235]: DEBUG nova.network.neutron [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.538139] env[62235]: DEBUG oslo_vmware.api [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271969, 'name': PowerOnVM_Task, 'duration_secs': 0.658267} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.538139] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.538139] env[62235]: INFO nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Took 7.79 seconds to spawn the instance on the hypervisor. [ 945.538635] env[62235]: DEBUG nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.540638] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8623d25-f92b-4512-88c4-f6c9569096bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.549541] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.549771] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.549953] env[62235]: DEBUG nova.network.neutron [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.680920] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271970, 'name': CreateVM_Task, 'duration_secs': 0.737381} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.681247] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.682272] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.682519] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.683014] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.683397] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee265769-bd13-4840-a9b6-d79f2f7d5076 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.688729] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 945.688729] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e69b88-db39-eef6-c3bb-7ba330b20475" [ 945.688729] env[62235]: _type = "Task" [ 945.688729] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.696382] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e69b88-db39-eef6-c3bb-7ba330b20475, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.846766] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.847026] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.847202] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.847397] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.847554] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.847829] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.848216] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.848438] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.848623] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.848818] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.849024] env[62235]: DEBUG nova.virt.hardware [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.850136] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7df8e1-5aed-4053-b238-04ba03798731 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.860659] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befd7507-e4c6-469e-9b94-16c1f505e51e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.873931] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:f2:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37fb1918-d178-4e12-93e6-316381e78be4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c536912e-8bb6-44c3-9ab4-fdc21a8c1556', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.882018] env[62235]: DEBUG oslo.service.loopingcall [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.882686] env[62235]: DEBUG oslo_concurrency.lockutils [req-0ac7045d-d0bf-4c57-a3b1-32f0bf89c919 req-6be13650-50eb-4cb4-8bd3-14b13876f25d service nova] Releasing lock "refresh_cache-79dafae1-e389-4e04-af20-577b8aac06a6" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.883057] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.883323] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-adc59905-0afc-44db-991d-3fc37327070f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.902543] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.902543] env[62235]: value = "task-1271973" [ 945.902543] env[62235]: _type = "Task" [ 945.902543] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.905615] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271972, 'name': ReconfigVM_Task, 'duration_secs': 0.272142} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.908466] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf/85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.909070] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af59587d-b320-4548-8d5d-43c392216e05 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.914949] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271973, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.916148] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 945.916148] env[62235]: value = "task-1271974" [ 945.916148] env[62235]: _type = "Task" [ 945.916148] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.924537] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271974, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.933491] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Applying migration context for instance 05aae40a-37ae-4bf1-91d0-02a7228fcff2 as it has an incoming, in-progress migration b1367deb-b42b-418c-b261-69119a3fb577. Migration status is pre-migrating {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 945.936356] env[62235]: INFO nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating resource usage from migration b1367deb-b42b-418c-b261-69119a3fb577 [ 945.967510] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3ea7db36-ebde-4523-b803-ad121a00269b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.967713] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 15d41c0f-045d-4e1c-88b5-1841f99f2e74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.967840] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0e9182f8-81e1-4ca2-97b6-0066faba0b22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.967985] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 062bbe32-3edd-4b06-b7d9-882e798b3a7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968124] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 2068403d-5870-4daa-b4d6-acaf39209488 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968242] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0ac4202d-4e25-417b-9ab0-6429b73e8e98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968358] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a3336b20-5737-4e5a-94a1-19fec140aa60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968514] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance eccde807-664e-4488-bec8-2d6919c2fc91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968581] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968767] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 79dafae1-e389-4e04-af20-577b8aac06a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968822] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance c47ea708-9a8b-44e6-a630-c20cfcd6f004 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.968924] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Migration b1367deb-b42b-418c-b261-69119a3fb577 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 945.969057] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 05aae40a-37ae-4bf1-91d0-02a7228fcff2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 946.063566] env[62235]: INFO nova.compute.manager [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Took 15.56 seconds to build instance. [ 946.107104] env[62235]: DEBUG nova.network.neutron [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.123180] env[62235]: DEBUG nova.compute.manager [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-changed-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.123724] env[62235]: DEBUG nova.compute.manager [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing instance network info cache due to event network-changed-64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.123724] env[62235]: DEBUG oslo_concurrency.lockutils [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.123890] env[62235]: DEBUG oslo_concurrency.lockutils [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.124012] env[62235]: DEBUG nova.network.neutron [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing network info cache for port 64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.198355] env[62235]: DEBUG nova.network.neutron [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance_info_cache with network_info: [{"id": "e3865625-b03d-42d2-8ba3-8829b7996727", "address": "fa:16:3e:2f:a8:fc", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3865625-b0", "ovs_interfaceid": "e3865625-b03d-42d2-8ba3-8829b7996727", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.207023] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e69b88-db39-eef6-c3bb-7ba330b20475, 'name': SearchDatastore_Task, 'duration_secs': 0.008831} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.207584] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.207827] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.208083] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.208434] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.208434] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.208701] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9b22018-e287-4219-9b53-cd32ed0655d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.218131] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.218331] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.219239] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f53272e-687f-480a-b1be-833d785db167 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.225398] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 946.225398] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e2bf1-d885-f2ef-2825-1bc3f718644c" [ 946.225398] env[62235]: _type = "Task" [ 946.225398] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.233765] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e2bf1-d885-f2ef-2825-1bc3f718644c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.311244] env[62235]: DEBUG nova.network.neutron [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Updating instance_info_cache with network_info: [{"id": "7d53cc57-246b-4fed-a55d-514cd12548d8", "address": "fa:16:3e:54:01:02", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d53cc57-24", "ovs_interfaceid": "7d53cc57-246b-4fed-a55d-514cd12548d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.415387] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271973, 'name': CreateVM_Task, 'duration_secs': 0.47624} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.415509] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.416179] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.416354] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.416669] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.416976] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21a3cd24-e93b-474e-a80a-430cce0181f5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.424246] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 946.424246] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52008af0-b298-fe5d-3ea6-dea961f632ce" [ 946.424246] env[62235]: _type = "Task" [ 946.424246] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.427640] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271974, 'name': Rename_Task, 'duration_secs': 0.154797} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.430742] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.431064] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38254468-4940-44e0-8928-ae777f93f8ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.437486] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52008af0-b298-fe5d-3ea6-dea961f632ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009892} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.438613] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.438850] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.439071] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.439338] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 946.439338] env[62235]: value = "task-1271975" [ 946.439338] env[62235]: _type = "Task" [ 946.439338] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.446209] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271975, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.472772] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 6f803e24-2b2b-4a15-8172-29acde2154f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 946.473039] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 946.473200] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3136MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 946.566209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8c765829-0526-4c0e-be4a-74350ad2463a tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.078s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.650946] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8a8486-c24d-4b71-9099-3e01d164d2c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.658950] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617f5a7f-a2cc-4715-96ea-1b026f9e1d85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.690726] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382b739d-cdd5-4957-a02b-503d658c822e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.699635] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e486772-5d57-4f27-bbc8-3b747fbb6333 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.714557] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.715584] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.741392] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e2bf1-d885-f2ef-2825-1bc3f718644c, 'name': SearchDatastore_Task, 'duration_secs': 0.009358} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.742307] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6244afda-9617-49ef-8c14-07cebf286bd3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.747247] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 946.747247] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f60644-096f-ba07-dca2-11609c7e15dd" [ 946.747247] env[62235]: _type = "Task" [ 946.747247] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.754983] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f60644-096f-ba07-dca2-11609c7e15dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.813767] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.814083] env[62235]: DEBUG nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Instance network_info: |[{"id": "7d53cc57-246b-4fed-a55d-514cd12548d8", "address": "fa:16:3e:54:01:02", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d53cc57-24", "ovs_interfaceid": "7d53cc57-246b-4fed-a55d-514cd12548d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.814506] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:01:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d53cc57-246b-4fed-a55d-514cd12548d8', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.822208] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Creating folder: Project (b690e5ef3dac4a07879b4b7fe2249e89). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.824613] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-541e21b7-cc0e-4c3f-b6d9-21cfc21c4e8b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.838672] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Created folder: Project (b690e5ef3dac4a07879b4b7fe2249e89) in parent group-v273362. [ 946.838795] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Creating folder: Instances. Parent ref: group-v273497. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.839183] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3cb9958-4003-46b1-96c1-3e4c0fe660a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.850995] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Created folder: Instances in parent group-v273497. [ 946.851278] env[62235]: DEBUG oslo.service.loopingcall [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.851479] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.851698] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-890afe1a-75cb-4699-8b0e-c98eca001e2b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.867758] env[62235]: DEBUG nova.network.neutron [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updated VIF entry in instance network info cache for port 64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 946.868168] env[62235]: DEBUG nova.network.neutron [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.874765] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.874765] env[62235]: value = "task-1271978" [ 946.874765] env[62235]: _type = "Task" [ 946.874765] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.888620] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271978, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.946089] env[62235]: DEBUG nova.compute.manager [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Received event network-changed-7d53cc57-246b-4fed-a55d-514cd12548d8 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.946477] env[62235]: DEBUG nova.compute.manager [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Refreshing instance network info cache due to event network-changed-7d53cc57-246b-4fed-a55d-514cd12548d8. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.946648] env[62235]: DEBUG oslo_concurrency.lockutils [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] Acquiring lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.946800] env[62235]: DEBUG oslo_concurrency.lockutils [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] Acquired lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.946965] env[62235]: DEBUG nova.network.neutron [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Refreshing network info cache for port 7d53cc57-246b-4fed-a55d-514cd12548d8 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.954689] env[62235]: DEBUG oslo_vmware.api [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271975, 'name': PowerOnVM_Task, 'duration_secs': 0.456036} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.955772] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.955772] env[62235]: INFO nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Took 6.93 seconds to spawn the instance on the hypervisor. [ 946.955916] env[62235]: DEBUG nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.956654] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538655bf-e3f9-47ea-b5f1-840009c36f68 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.222059] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.257337] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f60644-096f-ba07-dca2-11609c7e15dd, 'name': SearchDatastore_Task, 'duration_secs': 0.011088} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.258268] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.258537] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 79dafae1-e389-4e04-af20-577b8aac06a6/79dafae1-e389-4e04-af20-577b8aac06a6.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.258817] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.259027] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.259237] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e674b468-fb5b-467d-9ffe-9227571706ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.261267] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b2092a7-ecbf-489a-bd77-d87ec60b9436 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.267736] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 947.267736] env[62235]: value = "task-1271979" [ 947.267736] env[62235]: _type = "Task" [ 947.267736] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.271630] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.271805] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.272743] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c17fd555-44d5-452c-9c86-af11ac0b6510 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.278010] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271979, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.280779] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 947.280779] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232680d-2e27-633d-426e-bc80382941c6" [ 947.280779] env[62235]: _type = "Task" [ 947.280779] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.288248] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232680d-2e27-633d-426e-bc80382941c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.371363] env[62235]: DEBUG oslo_concurrency.lockutils [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.371729] env[62235]: DEBUG nova.compute.manager [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.371983] env[62235]: DEBUG nova.compute.manager [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing instance network info cache due to event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.372272] env[62235]: DEBUG oslo_concurrency.lockutils [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.372479] env[62235]: DEBUG oslo_concurrency.lockutils [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.372661] env[62235]: DEBUG nova.network.neutron [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.385281] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271978, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.484836] env[62235]: INFO nova.compute.manager [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Took 16.53 seconds to build instance. [ 947.673145] env[62235]: DEBUG nova.network.neutron [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Updated VIF entry in instance network info cache for port 7d53cc57-246b-4fed-a55d-514cd12548d8. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.674348] env[62235]: DEBUG nova.network.neutron [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Updating instance_info_cache with network_info: [{"id": "7d53cc57-246b-4fed-a55d-514cd12548d8", "address": "fa:16:3e:54:01:02", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d53cc57-24", "ovs_interfaceid": "7d53cc57-246b-4fed-a55d-514cd12548d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.728405] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 947.728729] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.816s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.729121] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.905s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.730614] env[62235]: INFO nova.compute.claims [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.778194] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271979, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495599} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.779157] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 79dafae1-e389-4e04-af20-577b8aac06a6/79dafae1-e389-4e04-af20-577b8aac06a6.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.779424] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.779720] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee863c69-e5be-416f-aaf4-689d86bc6d5e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.790894] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232680d-2e27-633d-426e-bc80382941c6, 'name': SearchDatastore_Task, 'duration_secs': 0.008962} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.792728] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 947.792728] env[62235]: value = "task-1271980" [ 947.792728] env[62235]: _type = "Task" [ 947.792728] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.792990] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0e353a9-5346-427e-9d09-1f7e150c0e84 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.803359] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 947.803359] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c28c29-2b2d-506c-5195-597efb37956d" [ 947.803359] env[62235]: _type = "Task" [ 947.803359] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.803648] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271980, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.810790] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c28c29-2b2d-506c-5195-597efb37956d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.887131] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1271978, 'name': CreateVM_Task, 'duration_secs': 0.688644} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.887313] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.888047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.888178] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.888445] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.888710] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43e83021-4743-433c-9da9-de479bbae8c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.893158] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 947.893158] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a6a472-1bc5-65b3-c99d-62e77a5aae93" [ 947.893158] env[62235]: _type = "Task" [ 947.893158] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.900605] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a6a472-1bc5-65b3-c99d-62e77a5aae93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.989309] env[62235]: DEBUG oslo_concurrency.lockutils [None req-414a6bd9-a245-4a0f-acec-0f36ae681e97 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.046s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.083514] env[62235]: DEBUG nova.network.neutron [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updated VIF entry in instance network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.083918] env[62235]: DEBUG nova.network.neutron [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.154878] env[62235]: DEBUG nova.compute.manager [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.155109] env[62235]: DEBUG nova.compute.manager [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing instance network info cache due to event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.155342] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.177705] env[62235]: DEBUG oslo_concurrency.lockutils [req-d61480db-df37-4a08-ac1f-668ae3554c7e req-f188086b-b772-40dc-ada9-6dfc4f491e15 service nova] Releasing lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.232970] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b7a07c-6a9d-4d3d-91a2-b43ecee001fb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.253196] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance '05aae40a-37ae-4bf1-91d0-02a7228fcff2' progress to 0 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 948.306921] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271980, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067958} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.310013] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.311185] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2015f8-bdf9-4e6d-bf71-db1179f4bea4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.318514] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c28c29-2b2d-506c-5195-597efb37956d, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.327543] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.327813] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.336680] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 79dafae1-e389-4e04-af20-577b8aac06a6/79dafae1-e389-4e04-af20-577b8aac06a6.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.336934] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-500449c8-b2dd-4d5c-b78b-1483ab3e6609 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.339027] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd3a52bb-7a0f-46a4-9ba6-ded6d2368d0b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.358189] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 948.358189] env[62235]: value = "task-1271982" [ 948.358189] env[62235]: _type = "Task" [ 948.358189] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.359521] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 948.359521] env[62235]: value = "task-1271981" [ 948.359521] env[62235]: _type = "Task" [ 948.359521] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.370940] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.374452] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271982, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.402526] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a6a472-1bc5-65b3-c99d-62e77a5aae93, 'name': SearchDatastore_Task, 'duration_secs': 0.009883} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.402829] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.403084] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.403323] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.403473] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.403654] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.403906] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bcb7e7b5-9715-4fd2-8bae-824211acf678 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.433794] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.434068] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.434749] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14b5e75d-3487-4bad-9109-3ed5f0914802 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.440246] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 948.440246] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a0824-960a-1f04-69c2-4722487f2b7a" [ 948.440246] env[62235]: _type = "Task" [ 948.440246] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.447726] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a0824-960a-1f04-69c2-4722487f2b7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.587315] env[62235]: DEBUG oslo_concurrency.lockutils [req-a99635d6-83c2-4381-a158-de5a1487808d req-3738b71f-a73e-446d-b2e3-17e12cf65951 service nova] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.587865] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.587992] env[62235]: DEBUG nova.network.neutron [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 948.758769] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.759113] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f832d1cd-af8f-40cd-a2ca-ea3d52ba1458 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.770092] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 948.770092] env[62235]: value = "task-1271983" [ 948.770092] env[62235]: _type = "Task" [ 948.770092] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.780515] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271983, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.872740] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271982, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.876859] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.927610] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf22f35-32ce-4467-b990-655551699c1c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.934182] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b614ad57-a34d-4753-b67e-b3b2adf69fb7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.971383] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e959ba-51fa-4e56-84bb-2f60895781aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.981752] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523a0824-960a-1f04-69c2-4722487f2b7a, 'name': SearchDatastore_Task, 'duration_secs': 0.055005} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.984907] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2189f802-6922-42ca-83f1-f90e093cc0ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.988406] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbe7894-e217-4f45-b649-9a60a36f174e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.998755] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 948.998755] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ce0a0f-9f07-b8cd-7361-d80e9b5ea3e0" [ 948.998755] env[62235]: _type = "Task" [ 948.998755] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.007405] env[62235]: DEBUG nova.compute.provider_tree [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 949.017659] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ce0a0f-9f07-b8cd-7361-d80e9b5ea3e0, 'name': SearchDatastore_Task, 'duration_secs': 0.015274} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.018614] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.018874] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/c47ea708-9a8b-44e6-a630-c20cfcd6f004.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.019196] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad985f3a-e1a0-4410-9f38-60795849458e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.025991] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 949.025991] env[62235]: value = "task-1271984" [ 949.025991] env[62235]: _type = "Task" [ 949.025991] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.034778] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.282550] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271983, 'name': PowerOffVM_Task, 'duration_secs': 0.259823} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.282851] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.283059] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance '05aae40a-37ae-4bf1-91d0-02a7228fcff2' progress to 17 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 949.372259] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271982, 'name': ReconfigVM_Task, 'duration_secs': 0.734721} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.375960] env[62235]: DEBUG nova.network.neutron [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updated VIF entry in instance network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 949.376349] env[62235]: DEBUG nova.network.neutron [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.377529] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 79dafae1-e389-4e04-af20-577b8aac06a6/79dafae1-e389-4e04-af20-577b8aac06a6.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.378182] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.996004} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.378581] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-799049be-ded0-46ee-a735-24eb0e52bdb9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.380152] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 949.380387] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.380632] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94b237b1-d834-4ead-9639-92f5a3cd9868 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.388280] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 949.388280] env[62235]: value = "task-1271986" [ 949.388280] env[62235]: _type = "Task" [ 949.388280] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.389806] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 949.389806] env[62235]: value = "task-1271985" [ 949.389806] env[62235]: _type = "Task" [ 949.389806] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.402446] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.405903] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271985, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.535839] env[62235]: ERROR nova.scheduler.client.report [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [req-4debaac6-b98b-4339-9665-149f18609f0b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4debaac6-b98b-4339-9665-149f18609f0b"}]} [ 949.543029] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271984, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.558446] env[62235]: DEBUG nova.scheduler.client.report [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 949.575860] env[62235]: DEBUG nova.scheduler.client.report [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 949.576217] env[62235]: DEBUG nova.compute.provider_tree [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 949.591099] env[62235]: DEBUG nova.scheduler.client.report [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 949.611412] env[62235]: DEBUG nova.scheduler.client.report [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 949.790203] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.790514] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.790625] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.790811] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.790963] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.791167] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.791385] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.791552] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.791724] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.791889] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.792113] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.799941] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d92bc68-e6f4-4fb1-af59-354381af1a57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.816600] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 949.816600] env[62235]: value = "task-1271987" [ 949.816600] env[62235]: _type = "Task" [ 949.816600] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.821648] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67d5536-98e8-40f3-ba4c-ee1e83f0a893 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.831353] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.832377] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce8c9c8-58ac-4d04-9261-be214887fb5f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.865018] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de3f9d1-c995-445b-b078-0d31d7e8b61e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.873172] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b564ecb-45a4-4e8e-aaad-eb277b37337f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.887017] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.887309] env[62235]: DEBUG nova.compute.manager [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-changed-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.887511] env[62235]: DEBUG nova.compute.manager [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing instance network info cache due to event network-changed-64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.887730] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.887906] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.888131] env[62235]: DEBUG nova.network.neutron [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing network info cache for port 64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.890189] env[62235]: DEBUG nova.compute.provider_tree [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 949.902673] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065941} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.905818] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.906430] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271985, 'name': Rename_Task, 'duration_secs': 0.18152} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.907162] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d32f08d-dbdc-4ae2-8e86-496b6322c6b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.909695] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.910205] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8838611-373e-447a-99d1-9328a7ec5103 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.937274] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.939237] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4b3c88e-f466-45ee-8ef6-60dbf42655ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.960027] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 949.960027] env[62235]: value = "task-1271988" [ 949.960027] env[62235]: _type = "Task" [ 949.960027] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.965143] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 949.965143] env[62235]: value = "task-1271989" [ 949.965143] env[62235]: _type = "Task" [ 949.965143] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.969178] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271988, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.978537] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271989, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.036942] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271984, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.701248} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.037237] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/c47ea708-9a8b-44e6-a630-c20cfcd6f004.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.037456] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.037716] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5ed0a46-94e7-4fef-b7c1-9fd4cef593d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.047893] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 950.047893] env[62235]: value = "task-1271990" [ 950.047893] env[62235]: _type = "Task" [ 950.047893] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.058150] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271990, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.327060] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271987, 'name': ReconfigVM_Task, 'duration_secs': 0.21273} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.327416] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance '05aae40a-37ae-4bf1-91d0-02a7228fcff2' progress to 33 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 950.422844] env[62235]: DEBUG nova.scheduler.client.report [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Updated inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with generation 106 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 950.423148] env[62235]: DEBUG nova.compute.provider_tree [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 106 to 107 during operation: update_inventory {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 950.423345] env[62235]: DEBUG nova.compute.provider_tree [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 950.470840] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271988, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.480383] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.560469] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271990, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087075} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.560769] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.561590] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e862acb-e539-479d-a9ca-8f5d82a6a9b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.585441] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/c47ea708-9a8b-44e6-a630-c20cfcd6f004.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.585765] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b28f5fbd-1d74-45c3-809f-7843612ce4b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.608048] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 950.608048] env[62235]: value = "task-1271991" [ 950.608048] env[62235]: _type = "Task" [ 950.608048] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.616185] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271991, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.635731] env[62235]: DEBUG nova.network.neutron [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updated VIF entry in instance network info cache for port 64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.636174] env[62235]: DEBUG nova.network.neutron [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.834263] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.834620] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.834697] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.834882] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.835047] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.835208] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.835431] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.835629] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.835809] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.835980] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.836178] env[62235]: DEBUG nova.virt.hardware [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.843552] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 950.843843] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e34f732-3997-4dd2-8e20-b98a7b3e7b4a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.863125] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 950.863125] env[62235]: value = "task-1271992" [ 950.863125] env[62235]: _type = "Task" [ 950.863125] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.875578] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271992, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.929979] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.201s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.930537] env[62235]: DEBUG nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 950.970743] env[62235]: DEBUG oslo_vmware.api [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271988, 'name': PowerOnVM_Task, 'duration_secs': 0.778876} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.974447] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 950.974685] env[62235]: INFO nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Took 8.65 seconds to spawn the instance on the hypervisor. [ 950.974873] env[62235]: DEBUG nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.976320] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90406aa-dd89-4539-a1e1-0d3d37d9e61f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.986704] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.118247] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271991, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.139313] env[62235]: DEBUG oslo_concurrency.lockutils [req-4b5f2b7e-fb32-49b1-ba95-a8d4af2ee906 req-96d4a334-b498-4ad9-a98d-e6561d45aaf6 service nova] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.374017] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271992, 'name': ReconfigVM_Task, 'duration_secs': 0.277309} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.374335] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 951.375152] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb727a86-fb3e-4086-a75a-012ddd76ee09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.397202] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 05aae40a-37ae-4bf1-91d0-02a7228fcff2/05aae40a-37ae-4bf1-91d0-02a7228fcff2.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.397491] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38332f47-e1d2-45e3-9a2a-8b98e4cf531d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.417465] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 951.417465] env[62235]: value = "task-1271993" [ 951.417465] env[62235]: _type = "Task" [ 951.417465] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.426103] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271993, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.435656] env[62235]: DEBUG nova.compute.utils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.437521] env[62235]: DEBUG nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.437652] env[62235]: DEBUG nova.network.neutron [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.480586] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.482091] env[62235]: DEBUG nova.policy [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96af7a114d04c0dad3c53f5a2dcd2bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef6547723ccf4be891da50a179dae06f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.497370] env[62235]: INFO nova.compute.manager [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Took 20.18 seconds to build instance. [ 951.619337] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271991, 'name': ReconfigVM_Task, 'duration_secs': 0.937064} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.619909] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Reconfigured VM instance instance-0000005a to attach disk [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/c47ea708-9a8b-44e6-a630-c20cfcd6f004.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.621741] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63b5a6bc-5bb5-474a-9103-d8f21c6f5384 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.630160] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 951.630160] env[62235]: value = "task-1271994" [ 951.630160] env[62235]: _type = "Task" [ 951.630160] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.643161] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271994, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.844106] env[62235]: DEBUG nova.network.neutron [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Successfully created port: 566501e6-2585-448d-a888-6a640269ed49 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.928336] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1271993, 'name': ReconfigVM_Task, 'duration_secs': 0.361577} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.928642] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 05aae40a-37ae-4bf1-91d0-02a7228fcff2/05aae40a-37ae-4bf1-91d0-02a7228fcff2.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.928919] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance '05aae40a-37ae-4bf1-91d0-02a7228fcff2' progress to 50 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 951.940388] env[62235]: DEBUG nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 951.985184] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271989, 'name': ReconfigVM_Task, 'duration_secs': 1.740163} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.986246] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 0ac4202d-4e25-417b-9ab0-6429b73e8e98/0ac4202d-4e25-417b-9ab0-6429b73e8e98.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.986246] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac483d15-e3ec-4c36-ac03-f3b48425877f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.994018] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 951.994018] env[62235]: value = "task-1271995" [ 951.994018] env[62235]: _type = "Task" [ 951.994018] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.998380] env[62235]: DEBUG oslo_concurrency.lockutils [None req-07a2b781-9e35-4eee-9272-4eadc5b9f196 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "79dafae1-e389-4e04-af20-577b8aac06a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.688s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.003276] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271995, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.141058] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271994, 'name': Rename_Task, 'duration_secs': 0.184856} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.141400] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.141652] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-127f9bd9-f904-4095-893b-c39745057e8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.149328] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 952.149328] env[62235]: value = "task-1271996" [ 952.149328] env[62235]: _type = "Task" [ 952.149328] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.157495] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.437035] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0944601d-b82d-463e-b672-400e3821a522 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.461210] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c792f084-d6f9-4ad6-901d-be53b4e1fe54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.481026] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance '05aae40a-37ae-4bf1-91d0-02a7228fcff2' progress to 67 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 952.503979] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271995, 'name': Rename_Task, 'duration_secs': 0.307569} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.504294] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.504604] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d252f7e1-c54f-401e-806b-5bf8f46a016a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.512259] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 952.512259] env[62235]: value = "task-1271997" [ 952.512259] env[62235]: _type = "Task" [ 952.512259] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.521019] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.661835] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271996, 'name': PowerOnVM_Task} progress is 74%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.965612] env[62235]: DEBUG nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 952.994288] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.994551] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.994716] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.994904] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.995074] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.995235] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.995520] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.995794] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.996057] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.996369] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.996458] env[62235]: DEBUG nova.virt.hardware [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.997311] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3da3cb3-3795-4b8a-9d16-7fce50597a3d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.006223] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e9361c-e9ac-4be7-a11a-c96ef6a505d0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.034819] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271997, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.035892] env[62235]: DEBUG nova.network.neutron [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Port e3865625-b03d-42d2-8ba3-8829b7996727 binding to destination host cpu-1 is already ACTIVE {{(pid=62235) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 953.160779] env[62235]: DEBUG oslo_vmware.api [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1271996, 'name': PowerOnVM_Task, 'duration_secs': 0.769771} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.161073] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.161317] env[62235]: INFO nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Took 8.55 seconds to spawn the instance on the hypervisor. [ 953.161613] env[62235]: DEBUG nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.162332] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818f4354-9501-4b31-aae0-95b47bce9781 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.259524] env[62235]: DEBUG oslo_concurrency.lockutils [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "eccde807-664e-4488-bec8-2d6919c2fc91" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.259860] env[62235]: DEBUG oslo_concurrency.lockutils [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.260099] env[62235]: DEBUG nova.compute.manager [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.261079] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04186c5b-c6eb-4081-9a1c-5871c3385c6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.269956] env[62235]: DEBUG nova.compute.manager [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62235) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 953.270663] env[62235]: DEBUG nova.objects.instance [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lazy-loading 'flavor' on Instance uuid eccde807-664e-4488-bec8-2d6919c2fc91 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.373554] env[62235]: DEBUG nova.compute.manager [req-a028f16b-51a2-4396-96ab-a1d652de8fc7 req-6c9ed4cf-a038-4f6d-93c0-05a614fe134b service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Received event network-vif-plugged-566501e6-2585-448d-a888-6a640269ed49 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.373878] env[62235]: DEBUG oslo_concurrency.lockutils [req-a028f16b-51a2-4396-96ab-a1d652de8fc7 req-6c9ed4cf-a038-4f6d-93c0-05a614fe134b service nova] Acquiring lock "6f803e24-2b2b-4a15-8172-29acde2154f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.374290] env[62235]: DEBUG oslo_concurrency.lockutils [req-a028f16b-51a2-4396-96ab-a1d652de8fc7 req-6c9ed4cf-a038-4f6d-93c0-05a614fe134b service nova] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.374566] env[62235]: DEBUG oslo_concurrency.lockutils [req-a028f16b-51a2-4396-96ab-a1d652de8fc7 req-6c9ed4cf-a038-4f6d-93c0-05a614fe134b service nova] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.374822] env[62235]: DEBUG nova.compute.manager [req-a028f16b-51a2-4396-96ab-a1d652de8fc7 req-6c9ed4cf-a038-4f6d-93c0-05a614fe134b service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] No waiting events found dispatching network-vif-plugged-566501e6-2585-448d-a888-6a640269ed49 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.375104] env[62235]: WARNING nova.compute.manager [req-a028f16b-51a2-4396-96ab-a1d652de8fc7 req-6c9ed4cf-a038-4f6d-93c0-05a614fe134b service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Received unexpected event network-vif-plugged-566501e6-2585-448d-a888-6a640269ed49 for instance with vm_state building and task_state spawning. [ 953.464629] env[62235]: DEBUG nova.network.neutron [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Successfully updated port: 566501e6-2585-448d-a888-6a640269ed49 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.523947] env[62235]: DEBUG oslo_vmware.api [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271997, 'name': PowerOnVM_Task, 'duration_secs': 0.571538} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.524269] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.524484] env[62235]: DEBUG nova.compute.manager [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.525291] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4aa4518-ec64-4f73-9940-8781ee370caf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.680475] env[62235]: INFO nova.compute.manager [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Took 21.16 seconds to build instance. [ 953.776338] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.776338] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5426b943-34df-432b-a8a4-afa228cca139 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.785965] env[62235]: DEBUG oslo_vmware.api [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 953.785965] env[62235]: value = "task-1271998" [ 953.785965] env[62235]: _type = "Task" [ 953.785965] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.794946] env[62235]: DEBUG oslo_vmware.api [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.966953] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-6f803e24-2b2b-4a15-8172-29acde2154f8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.967491] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-6f803e24-2b2b-4a15-8172-29acde2154f8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.967491] env[62235]: DEBUG nova.network.neutron [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.049518] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.049777] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.050032] env[62235]: DEBUG nova.objects.instance [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 954.059155] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.059155] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.059155] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.183766] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ef260596-9100-4ac8-b496-45b90bb150a9 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.673s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.297586] env[62235]: DEBUG oslo_vmware.api [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1271998, 'name': PowerOffVM_Task, 'duration_secs': 0.294178} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.297872] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.298070] env[62235]: DEBUG nova.compute.manager [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.298850] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80f1e8a-2508-49de-b74c-f7c4d70d5f8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.499015] env[62235]: DEBUG nova.network.neutron [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.644060] env[62235]: DEBUG nova.network.neutron [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Updating instance_info_cache with network_info: [{"id": "566501e6-2585-448d-a888-6a640269ed49", "address": "fa:16:3e:a4:08:f8", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap566501e6-25", "ovs_interfaceid": "566501e6-2585-448d-a888-6a640269ed49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.785686] env[62235]: INFO nova.compute.manager [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Rescuing [ 954.786088] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.786333] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.786594] env[62235]: DEBUG nova.network.neutron [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.813512] env[62235]: DEBUG oslo_concurrency.lockutils [None req-94d358b7-3d3b-4045-bcd0-cd09fc8feabb tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.553s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.067986] env[62235]: DEBUG oslo_concurrency.lockutils [None req-165d9865-b177-45b5-a3b4-b0e6ef9bd8bf tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.075375] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.075656] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.075967] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.076094] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.076245] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.078666] env[62235]: INFO nova.compute.manager [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Terminating instance [ 955.080966] env[62235]: DEBUG nova.compute.manager [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.081179] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.082068] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a580c0d-eb5a-496d-b3c3-6b88b052974f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.091586] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.091867] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2fa1651f-02ce-4203-aa5a-beb122404d01 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.098742] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.098931] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.099227] env[62235]: DEBUG nova.network.neutron [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.103045] env[62235]: DEBUG oslo_vmware.api [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 955.103045] env[62235]: value = "task-1271999" [ 955.103045] env[62235]: _type = "Task" [ 955.103045] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.113972] env[62235]: DEBUG oslo_vmware.api [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271999, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.146716] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-6f803e24-2b2b-4a15-8172-29acde2154f8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.147084] env[62235]: DEBUG nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Instance network_info: |[{"id": "566501e6-2585-448d-a888-6a640269ed49", "address": "fa:16:3e:a4:08:f8", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap566501e6-25", "ovs_interfaceid": "566501e6-2585-448d-a888-6a640269ed49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.147535] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:08:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '566501e6-2585-448d-a888-6a640269ed49', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.155589] env[62235]: DEBUG oslo.service.loopingcall [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.156209] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 955.156453] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bed979c-3cde-4d91-b591-9ed6f6186f16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.180299] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.180299] env[62235]: value = "task-1272000" [ 955.180299] env[62235]: _type = "Task" [ 955.180299] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.190075] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272000, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.402455] env[62235]: DEBUG nova.compute.manager [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Received event network-changed-566501e6-2585-448d-a888-6a640269ed49 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.402659] env[62235]: DEBUG nova.compute.manager [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Refreshing instance network info cache due to event network-changed-566501e6-2585-448d-a888-6a640269ed49. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.402913] env[62235]: DEBUG oslo_concurrency.lockutils [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] Acquiring lock "refresh_cache-6f803e24-2b2b-4a15-8172-29acde2154f8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.403039] env[62235]: DEBUG oslo_concurrency.lockutils [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] Acquired lock "refresh_cache-6f803e24-2b2b-4a15-8172-29acde2154f8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.403206] env[62235]: DEBUG nova.network.neutron [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Refreshing network info cache for port 566501e6-2585-448d-a888-6a640269ed49 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.583343] env[62235]: DEBUG nova.network.neutron [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Updating instance_info_cache with network_info: [{"id": "7d53cc57-246b-4fed-a55d-514cd12548d8", "address": "fa:16:3e:54:01:02", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d53cc57-24", "ovs_interfaceid": "7d53cc57-246b-4fed-a55d-514cd12548d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.615611] env[62235]: DEBUG oslo_vmware.api [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1271999, 'name': PowerOffVM_Task, 'duration_secs': 0.241702} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.615888] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.616071] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.616335] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53991ec3-b0dc-4e98-82ab-236ab6864368 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.690602] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272000, 'name': CreateVM_Task, 'duration_secs': 0.374486} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.693579] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 955.694879] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.695077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.695414] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.696498] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.696498] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.696498] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleting the datastore file [datastore1] 0ac4202d-4e25-417b-9ab0-6429b73e8e98 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.696672] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d303b34-e64f-485e-bb97-d7ac94ab5d5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.698419] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e89ce625-8027-47e1-8e48-7707a5cc7454 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.703333] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 955.703333] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fcd896-d96f-2a07-8578-59624614473b" [ 955.703333] env[62235]: _type = "Task" [ 955.703333] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.707970] env[62235]: DEBUG oslo_vmware.api [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 955.707970] env[62235]: value = "task-1272002" [ 955.707970] env[62235]: _type = "Task" [ 955.707970] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.714439] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fcd896-d96f-2a07-8578-59624614473b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.722842] env[62235]: DEBUG oslo_vmware.api [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272002, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.731216] env[62235]: DEBUG nova.objects.instance [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lazy-loading 'flavor' on Instance uuid eccde807-664e-4488-bec8-2d6919c2fc91 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.815834] env[62235]: DEBUG nova.network.neutron [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance_info_cache with network_info: [{"id": "e3865625-b03d-42d2-8ba3-8829b7996727", "address": "fa:16:3e:2f:a8:fc", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3865625-b0", "ovs_interfaceid": "e3865625-b03d-42d2-8ba3-8829b7996727", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.085818] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "refresh_cache-c47ea708-9a8b-44e6-a630-c20cfcd6f004" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.154959] env[62235]: DEBUG nova.network.neutron [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Updated VIF entry in instance network info cache for port 566501e6-2585-448d-a888-6a640269ed49. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.155343] env[62235]: DEBUG nova.network.neutron [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Updating instance_info_cache with network_info: [{"id": "566501e6-2585-448d-a888-6a640269ed49", "address": "fa:16:3e:a4:08:f8", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap566501e6-25", "ovs_interfaceid": "566501e6-2585-448d-a888-6a640269ed49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.215204] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52fcd896-d96f-2a07-8578-59624614473b, 'name': SearchDatastore_Task, 'duration_secs': 0.011174} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.215828] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.216086] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.216324] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.216478] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.216664] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.216922] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0156e5ab-e93f-4d4c-9bb9-804f503a1690 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.221279] env[62235]: DEBUG oslo_vmware.api [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272002, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182203} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.221817] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.222015] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.222210] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.222393] env[62235]: INFO nova.compute.manager [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Took 1.14 seconds to destroy the instance on the hypervisor. [ 956.222627] env[62235]: DEBUG oslo.service.loopingcall [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.222876] env[62235]: DEBUG nova.compute.manager [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.222933] env[62235]: DEBUG nova.network.neutron [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.227709] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.227884] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 956.228578] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd2afcbd-3260-4f78-8956-810df2832f3b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.235915] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 956.235915] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ced20e-bb9c-cb5a-7e6c-b95305cc94d3" [ 956.235915] env[62235]: _type = "Task" [ 956.235915] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.236448] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.236618] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquired lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.236794] env[62235]: DEBUG nova.network.neutron [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.236981] env[62235]: DEBUG nova.objects.instance [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lazy-loading 'info_cache' on Instance uuid eccde807-664e-4488-bec8-2d6919c2fc91 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.246269] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ced20e-bb9c-cb5a-7e6c-b95305cc94d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008161} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.247072] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95878565-168f-44b9-8840-1e596d7f7703 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.252586] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 956.252586] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b20a9e-2a44-1850-e96f-caf7da1ace73" [ 956.252586] env[62235]: _type = "Task" [ 956.252586] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.262909] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b20a9e-2a44-1850-e96f-caf7da1ace73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.318898] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.623732] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.624033] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-325d0f6b-dfcd-45c1-b11e-65c54f6b5f95 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.633025] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 956.633025] env[62235]: value = "task-1272003" [ 956.633025] env[62235]: _type = "Task" [ 956.633025] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.642792] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.657625] env[62235]: DEBUG oslo_concurrency.lockutils [req-57356775-6d13-4da8-abbc-3a040dab32d8 req-00d6dd97-a6ea-4a56-a988-708b5cded7d3 service nova] Releasing lock "refresh_cache-6f803e24-2b2b-4a15-8172-29acde2154f8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.743154] env[62235]: DEBUG nova.objects.base [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 956.764501] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b20a9e-2a44-1850-e96f-caf7da1ace73, 'name': SearchDatastore_Task, 'duration_secs': 0.013564} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.764776] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.765057] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 6f803e24-2b2b-4a15-8172-29acde2154f8/6f803e24-2b2b-4a15-8172-29acde2154f8.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 956.765324] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7b7510e-c6ee-452a-a749-69a94dc234c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.773337] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 956.773337] env[62235]: value = "task-1272004" [ 956.773337] env[62235]: _type = "Task" [ 956.773337] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.787521] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.844269] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f72ee3-3865-41ae-88b8-5350f557aa75 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.865437] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afe7e42-251d-4b65-b60e-669dd4778880 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.873009] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance '05aae40a-37ae-4bf1-91d0-02a7228fcff2' progress to 83 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 956.937204] env[62235]: DEBUG nova.network.neutron [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.145673] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272003, 'name': PowerOffVM_Task, 'duration_secs': 0.195062} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.146056] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.147031] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55aa579d-fbb6-49f7-b74d-de62d8aa2817 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.169919] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ecc654-29f4-4a4c-abab-1cc8f21fb037 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.207294] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.207822] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec34f41b-f34a-46c4-a427-3644f591840a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.217666] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 957.217666] env[62235]: value = "task-1272005" [ 957.217666] env[62235]: _type = "Task" [ 957.217666] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.227741] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.285760] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272004, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.381105] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.381500] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-2068403d-5870-4daa-b4d6-acaf39209488-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.381728] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-2068403d-5870-4daa-b4d6-acaf39209488-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.382051] env[62235]: DEBUG nova.objects.instance [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'flavor' on Instance uuid 2068403d-5870-4daa-b4d6-acaf39209488 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.383256] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e708e6a5-c4b5-4747-a06f-48b34ae0a877 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.390987] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 957.390987] env[62235]: value = "task-1272006" [ 957.390987] env[62235]: _type = "Task" [ 957.390987] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.399882] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.425602] env[62235]: DEBUG nova.compute.manager [req-612c7701-cc9e-45df-ba8f-e1db8dd6c8d9 req-1869fa82-6a40-4457-9543-815b5803d4d4 service nova] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Received event network-vif-deleted-c536912e-8bb6-44c3-9ab4-fdc21a8c1556 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.439546] env[62235]: INFO nova.compute.manager [-] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Took 1.22 seconds to deallocate network for instance. [ 957.450888] env[62235]: DEBUG nova.network.neutron [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Updating instance_info_cache with network_info: [{"id": "f505e074-0353-4a28-abad-d058112f253a", "address": "fa:16:3e:20:74:13", "network": {"id": "1b44db11-2ef8-478e-b1df-5ebc938e1c25", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-950018915-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb228bc090c468d833754afe2bcb3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf505e074-03", "ovs_interfaceid": "f505e074-0353-4a28-abad-d058112f253a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.736523] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 957.736855] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.737197] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.737432] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.737575] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.738228] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8939d338-65ab-43e3-9349-ec13168812c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.750777] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.750777] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.751910] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62f2bae9-4d98-4b68-b7ca-5f430ff3352c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.758777] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 957.758777] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5246d7bf-04d0-6f51-fea7-e1d28ce19b70" [ 957.758777] env[62235]: _type = "Task" [ 957.758777] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.766687] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5246d7bf-04d0-6f51-fea7-e1d28ce19b70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.783369] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272004, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51898} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.783645] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 6f803e24-2b2b-4a15-8172-29acde2154f8/6f803e24-2b2b-4a15-8172-29acde2154f8.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 957.783950] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 957.784222] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4779cadc-7457-4666-b43a-c5372ebcca94 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.791336] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 957.791336] env[62235]: value = "task-1272007" [ 957.791336] env[62235]: _type = "Task" [ 957.791336] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.800090] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272007, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.901070] env[62235]: DEBUG oslo_vmware.api [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272006, 'name': PowerOnVM_Task, 'duration_secs': 0.369238} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.901400] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.901610] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1088691b-0a81-40da-954c-f980316dd4e5 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance '05aae40a-37ae-4bf1-91d0-02a7228fcff2' progress to 100 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 957.946777] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.947086] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.947324] env[62235]: DEBUG nova.objects.instance [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'resources' on Instance uuid 0ac4202d-4e25-417b-9ab0-6429b73e8e98 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.953967] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Releasing lock "refresh_cache-eccde807-664e-4488-bec8-2d6919c2fc91" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.986134] env[62235]: DEBUG nova.objects.instance [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'pci_requests' on Instance uuid 2068403d-5870-4daa-b4d6-acaf39209488 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.270064] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5246d7bf-04d0-6f51-fea7-e1d28ce19b70, 'name': SearchDatastore_Task, 'duration_secs': 0.009607} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.270471] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b39d9ce2-adec-4609-8d8d-32fd63d7269a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.296956] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 958.296956] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529cf49b-8100-f41a-5c9b-0904427ff113" [ 958.296956] env[62235]: _type = "Task" [ 958.296956] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.303450] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272007, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071227} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.304037] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 958.304816] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f22aa6-dc10-4441-bc81-ba635e26f0df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.311980] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529cf49b-8100-f41a-5c9b-0904427ff113, 'name': SearchDatastore_Task, 'duration_secs': 0.010627} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.312575] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.312857] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk. {{(pid=62235) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 958.313138] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b1edce1-c39f-45dc-964a-a214c48f18ca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.333619] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 6f803e24-2b2b-4a15-8172-29acde2154f8/6f803e24-2b2b-4a15-8172-29acde2154f8.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.334304] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10feaa0e-6c82-4205-a55e-db43de53d066 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.350649] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 958.350649] env[62235]: value = "task-1272008" [ 958.350649] env[62235]: _type = "Task" [ 958.350649] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.357403] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 958.357403] env[62235]: value = "task-1272009" [ 958.357403] env[62235]: _type = "Task" [ 958.357403] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.360537] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.369071] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272009, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.456242] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.456242] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a1dedbc-eb25-41a1-9387-8d7d96e1d4e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.465453] env[62235]: DEBUG oslo_vmware.api [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 958.465453] env[62235]: value = "task-1272010" [ 958.465453] env[62235]: _type = "Task" [ 958.465453] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.473377] env[62235]: DEBUG oslo_vmware.api [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272010, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.488416] env[62235]: DEBUG nova.objects.base [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Object Instance<2068403d-5870-4daa-b4d6-acaf39209488> lazy-loaded attributes: flavor,pci_requests {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 958.488677] env[62235]: DEBUG nova.network.neutron [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 958.547359] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.547689] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.547835] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.548020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.548244] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.550615] env[62235]: INFO nova.compute.manager [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Terminating instance [ 958.553301] env[62235]: DEBUG nova.compute.manager [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.553561] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.554531] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45bf1d59-1e45-453c-81e0-c7c0b8389370 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.566356] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.566725] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eeba0cd0-d31e-4f46-9952-e23fd4c285c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.575887] env[62235]: DEBUG nova.policy [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.580161] env[62235]: DEBUG oslo_vmware.api [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 958.580161] env[62235]: value = "task-1272011" [ 958.580161] env[62235]: _type = "Task" [ 958.580161] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.591381] env[62235]: DEBUG oslo_vmware.api [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272011, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.679286] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd86f28-7ebd-4302-a27e-8add461194bf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.691519] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21b47a5-8d53-4228-9888-22730b42ed01 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.728856] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b66f2cc-0642-4567-b048-24a97938ca3c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.740318] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdab60df-639b-4c10-82cf-2306c8e05d84 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.759966] env[62235]: DEBUG nova.compute.provider_tree [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.864156] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522292} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.867753] env[62235]: INFO nova.virt.vmwareapi.ds_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk. [ 958.868657] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea9a2e6-3c88-4b3d-995d-ac65043a6b7c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.879254] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272009, 'name': ReconfigVM_Task, 'duration_secs': 0.519815} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.892423] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 6f803e24-2b2b-4a15-8172-29acde2154f8/6f803e24-2b2b-4a15-8172-29acde2154f8.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.899979] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.900259] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f11db36-5ad2-4032-98df-4d55f24334f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.902035] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e735f62d-a905-473e-aca2-382bfdbe7c6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.926266] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 958.926266] env[62235]: value = "task-1272013" [ 958.926266] env[62235]: _type = "Task" [ 958.926266] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.927679] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 958.927679] env[62235]: value = "task-1272012" [ 958.927679] env[62235]: _type = "Task" [ 958.927679] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.947202] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272013, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.947381] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272012, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.975041] env[62235]: DEBUG oslo_vmware.api [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272010, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.090983] env[62235]: DEBUG oslo_vmware.api [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272011, 'name': PowerOffVM_Task, 'duration_secs': 0.243643} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.091325] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.091529] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.091804] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d71b9327-39a1-4193-85b8-9be31dff579a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.263821] env[62235]: DEBUG nova.scheduler.client.report [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.443425] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272013, 'name': Rename_Task, 'duration_secs': 0.149435} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.447271] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 959.447546] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272012, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.447772] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1a1599c-0e90-4efb-91f0-40a6a66fec86 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.457525] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 959.457525] env[62235]: value = "task-1272015" [ 959.457525] env[62235]: _type = "Task" [ 959.457525] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.473609] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272015, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.478920] env[62235]: DEBUG oslo_vmware.api [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272010, 'name': PowerOnVM_Task, 'duration_secs': 0.622033} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.479210] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.479405] env[62235]: DEBUG nova.compute.manager [None req-d53f70f9-5f7a-4ece-b952-26182cef4bce tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.480183] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae46df5-1744-400e-8c74-0fb38b6e0591 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.768417] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.821s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.790963] env[62235]: INFO nova.scheduler.client.report [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted allocations for instance 0ac4202d-4e25-417b-9ab0-6429b73e8e98 [ 959.945414] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272012, 'name': ReconfigVM_Task, 'duration_secs': 0.807376} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.946615] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Reconfigured VM instance instance-0000005a to attach disk [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.947283] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb0e012-a863-4d88-9dfe-5193207b1318 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.981009] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-708bc1ab-1cb4-4ab8-afb4-cb7f51fc01cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.001211] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272015, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.002807] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 960.002807] env[62235]: value = "task-1272016" [ 960.002807] env[62235]: _type = "Task" [ 960.002807] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.013600] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272016, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.051908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.052225] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.052316] env[62235]: DEBUG nova.compute.manager [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Going to confirm migration 2 {{(pid=62235) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 960.128431] env[62235]: DEBUG nova.compute.manager [req-edc493e9-5088-4eab-80ec-39ffc977965e req-3120da17-aef3-459e-8044-a29f9946f3ec service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-vif-plugged-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.128714] env[62235]: DEBUG oslo_concurrency.lockutils [req-edc493e9-5088-4eab-80ec-39ffc977965e req-3120da17-aef3-459e-8044-a29f9946f3ec service nova] Acquiring lock "2068403d-5870-4daa-b4d6-acaf39209488-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.128988] env[62235]: DEBUG oslo_concurrency.lockutils [req-edc493e9-5088-4eab-80ec-39ffc977965e req-3120da17-aef3-459e-8044-a29f9946f3ec service nova] Lock "2068403d-5870-4daa-b4d6-acaf39209488-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.129245] env[62235]: DEBUG oslo_concurrency.lockutils [req-edc493e9-5088-4eab-80ec-39ffc977965e req-3120da17-aef3-459e-8044-a29f9946f3ec service nova] Lock "2068403d-5870-4daa-b4d6-acaf39209488-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.129457] env[62235]: DEBUG nova.compute.manager [req-edc493e9-5088-4eab-80ec-39ffc977965e req-3120da17-aef3-459e-8044-a29f9946f3ec service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] No waiting events found dispatching network-vif-plugged-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.129664] env[62235]: WARNING nova.compute.manager [req-edc493e9-5088-4eab-80ec-39ffc977965e req-3120da17-aef3-459e-8044-a29f9946f3ec service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received unexpected event network-vif-plugged-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b for instance with vm_state active and task_state None. [ 960.143419] env[62235]: DEBUG nova.network.neutron [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Successfully updated port: ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.186896] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 960.187154] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 960.187432] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleting the datastore file [datastore1] 062bbe32-3edd-4b06-b7d9-882e798b3a7f {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.187754] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6bd06ba8-ca41-43b9-a928-19f4b5d0c480 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.197737] env[62235]: DEBUG oslo_vmware.api [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 960.197737] env[62235]: value = "task-1272017" [ 960.197737] env[62235]: _type = "Task" [ 960.197737] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.207242] env[62235]: DEBUG oslo_vmware.api [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.299604] env[62235]: DEBUG oslo_concurrency.lockutils [None req-fd2f93e7-e65f-45c0-8e45-8fcb17e7d8b4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "0ac4202d-4e25-417b-9ab0-6429b73e8e98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.224s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.485694] env[62235]: DEBUG oslo_vmware.api [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272015, 'name': PowerOnVM_Task, 'duration_secs': 0.675982} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.486518] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.486945] env[62235]: INFO nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Took 7.52 seconds to spawn the instance on the hypervisor. [ 960.487307] env[62235]: DEBUG nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.489077] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e36ce94-5900-41b0-b20a-3afbff0a19ec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.513028] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272016, 'name': ReconfigVM_Task, 'duration_secs': 0.226137} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.513028] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.513226] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a85c1335-5613-4799-a656-bdbc2abfd7c6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.520777] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 960.520777] env[62235]: value = "task-1272018" [ 960.520777] env[62235]: _type = "Task" [ 960.520777] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.529446] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272018, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.620124] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.620329] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.620514] env[62235]: DEBUG nova.network.neutron [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.620705] env[62235]: DEBUG nova.objects.instance [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lazy-loading 'info_cache' on Instance uuid 05aae40a-37ae-4bf1-91d0-02a7228fcff2 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.646102] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.646369] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.646561] env[62235]: DEBUG nova.network.neutron [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.708473] env[62235]: DEBUG oslo_vmware.api [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215673} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.708751] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.708943] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 960.709173] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 960.709360] env[62235]: INFO nova.compute.manager [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Took 2.16 seconds to destroy the instance on the hypervisor. [ 960.709642] env[62235]: DEBUG oslo.service.loopingcall [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.709853] env[62235]: DEBUG nova.compute.manager [-] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 960.709947] env[62235]: DEBUG nova.network.neutron [-] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 961.008582] env[62235]: INFO nova.compute.manager [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Took 27.21 seconds to build instance. [ 961.037121] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272018, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.190358] env[62235]: WARNING nova.network.neutron [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] 695860c9-9102-4053-ad1c-75a1c4ac4b12 already exists in list: networks containing: ['695860c9-9102-4053-ad1c-75a1c4ac4b12']. ignoring it [ 961.509523] env[62235]: DEBUG oslo_concurrency.lockutils [None req-90fac288-f4a4-41fa-ae36-250798610d22 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.720s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.519849] env[62235]: DEBUG nova.network.neutron [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "address": "fa:16:3e:77:07:e1", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac6a259e-eb", "ovs_interfaceid": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.533094] env[62235]: DEBUG oslo_vmware.api [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272018, 'name': PowerOnVM_Task, 'duration_secs': 0.549887} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.534011] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 961.536583] env[62235]: DEBUG nova.compute.manager [None req-fd38c9dc-6b10-4301-a5bb-7b93920ad0bf tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.537386] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9008596-2345-4ef7-80fa-ce3bf0bab1c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.680741] env[62235]: DEBUG nova.network.neutron [-] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.906911] env[62235]: DEBUG nova.network.neutron [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance_info_cache with network_info: [{"id": "e3865625-b03d-42d2-8ba3-8829b7996727", "address": "fa:16:3e:2f:a8:fc", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3865625-b0", "ovs_interfaceid": "e3865625-b03d-42d2-8ba3-8829b7996727", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.023214] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.023964] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.024150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.025245] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb52cd5-9652-4894-9920-84e0924c9429 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.044670] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.044916] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.045102] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.045301] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.045453] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.045606] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.045847] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.046031] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.046212] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.046382] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.046559] env[62235]: DEBUG nova.virt.hardware [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.052778] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Reconfiguring VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 962.055806] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a41ff09b-3172-4899-8b1b-b2a3c03e4d1b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.076020] env[62235]: DEBUG oslo_vmware.api [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 962.076020] env[62235]: value = "task-1272019" [ 962.076020] env[62235]: _type = "Task" [ 962.076020] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.086179] env[62235]: DEBUG oslo_vmware.api [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272019, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.183785] env[62235]: INFO nova.compute.manager [-] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Took 1.47 seconds to deallocate network for instance. [ 962.410523] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-05aae40a-37ae-4bf1-91d0-02a7228fcff2" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.410814] env[62235]: DEBUG nova.objects.instance [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lazy-loading 'migration_context' on Instance uuid 05aae40a-37ae-4bf1-91d0-02a7228fcff2 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.500420] env[62235]: DEBUG nova.compute.manager [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-changed-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.500664] env[62235]: DEBUG nova.compute.manager [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing instance network info cache due to event network-changed-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.500912] env[62235]: DEBUG oslo_concurrency.lockutils [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.501040] env[62235]: DEBUG oslo_concurrency.lockutils [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.501210] env[62235]: DEBUG nova.network.neutron [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing network info cache for port ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.589638] env[62235]: DEBUG oslo_vmware.api [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272019, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.690797] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.691168] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.691463] env[62235]: DEBUG nova.objects.instance [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'resources' on Instance uuid 062bbe32-3edd-4b06-b7d9-882e798b3a7f {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.888103] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "6f803e24-2b2b-4a15-8172-29acde2154f8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.888409] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.888604] env[62235]: DEBUG nova.compute.manager [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.889560] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f7eee1-c52c-46a4-ae0c-abb9e14a0cd1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.896930] env[62235]: DEBUG nova.compute.manager [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62235) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 962.897735] env[62235]: DEBUG nova.objects.instance [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'flavor' on Instance uuid 6f803e24-2b2b-4a15-8172-29acde2154f8 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.914072] env[62235]: DEBUG nova.objects.base [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Object Instance<05aae40a-37ae-4bf1-91d0-02a7228fcff2> lazy-loaded attributes: info_cache,migration_context {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 962.916114] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4beb3e-04d5-4aa0-ab88-70f36741b787 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.940701] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af6ad963-bfb5-4875-9f2d-f0c654b4cb06 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.946914] env[62235]: DEBUG oslo_vmware.api [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 962.946914] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5214ffb5-fba2-8353-7f17-b6513b0dc39b" [ 962.946914] env[62235]: _type = "Task" [ 962.946914] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.955530] env[62235]: DEBUG oslo_vmware.api [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5214ffb5-fba2-8353-7f17-b6513b0dc39b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.087567] env[62235]: DEBUG oslo_vmware.api [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272019, 'name': ReconfigVM_Task, 'duration_secs': 0.768439} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.088140] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.088385] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Reconfigured VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 963.273710] env[62235]: DEBUG nova.compute.manager [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Stashing vm_state: active {{(pid=62235) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 963.308827] env[62235]: DEBUG nova.network.neutron [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updated VIF entry in instance network info cache for port ac6a259e-ebe5-4b85-b07f-38fe6b16c73b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 963.309269] env[62235]: DEBUG nova.network.neutron [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "address": "fa:16:3e:77:07:e1", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac6a259e-eb", "ovs_interfaceid": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.390701] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172059af-a80a-48ce-8675-c648d65bcb30 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.401967] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a8e0fa-96a5-4d10-86c2-e2b53d268e48 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.405436] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.405968] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47864e3d-65fc-41f1-9bb9-30b88fff963b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.436843] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1e7f41-9c1e-4ebe-bc14-80abfcb309e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.439610] env[62235]: DEBUG oslo_vmware.api [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 963.439610] env[62235]: value = "task-1272020" [ 963.439610] env[62235]: _type = "Task" [ 963.439610] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.456220] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809dd386-00a6-4b45-bac8-35a62e872948 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.460210] env[62235]: DEBUG oslo_vmware.api [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.466806] env[62235]: DEBUG oslo_vmware.api [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5214ffb5-fba2-8353-7f17-b6513b0dc39b, 'name': SearchDatastore_Task, 'duration_secs': 0.008958} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.474812] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.475857] env[62235]: DEBUG nova.compute.provider_tree [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.593995] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b82997ea-179d-47eb-b4b0-77821ac15815 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-2068403d-5870-4daa-b4d6-acaf39209488-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.212s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.791561] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.812975] env[62235]: DEBUG oslo_concurrency.lockutils [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.812975] env[62235]: DEBUG nova.compute.manager [req-ed3e3523-e24c-4c65-a602-55c35f62a586 req-bde1589e-ad51-440f-8bb7-269ef02f229d service nova] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Received event network-vif-deleted-984bbc0d-59a6-46fc-bb73-bcc8454e14e3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.950777] env[62235]: DEBUG oslo_vmware.api [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272020, 'name': PowerOffVM_Task, 'duration_secs': 0.296919} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.951075] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 963.951259] env[62235]: DEBUG nova.compute.manager [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.952078] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313fef52-69dc-4093-b1d7-02e44caae951 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.979645] env[62235]: DEBUG nova.scheduler.client.report [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.040953] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "550945fe-52be-408d-9f42-f5cd5e9c489f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.041108] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.466347] env[62235]: DEBUG oslo_concurrency.lockutils [None req-7ef026dc-d11e-44f5-b5ee-af0711e98086 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.578s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.486021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.793s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.488304] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.013s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.509081] env[62235]: INFO nova.scheduler.client.report [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted allocations for instance 062bbe32-3edd-4b06-b7d9-882e798b3a7f [ 964.545531] env[62235]: DEBUG nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.670080] env[62235]: DEBUG oslo_concurrency.lockutils [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.670080] env[62235]: DEBUG oslo_concurrency.lockutils [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.831402] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "79dafae1-e389-4e04-af20-577b8aac06a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.831725] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "79dafae1-e389-4e04-af20-577b8aac06a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.831923] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "79dafae1-e389-4e04-af20-577b8aac06a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.832132] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "79dafae1-e389-4e04-af20-577b8aac06a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.832312] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "79dafae1-e389-4e04-af20-577b8aac06a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.834469] env[62235]: INFO nova.compute.manager [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Terminating instance [ 964.838036] env[62235]: DEBUG nova.compute.manager [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.838248] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.839120] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991af2ef-782e-4c05-84fa-94f38f873f87 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.854266] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.855037] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce647743-3c63-4c66-9c86-a7a56cdbbda9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.863839] env[62235]: DEBUG oslo_vmware.api [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 964.863839] env[62235]: value = "task-1272021" [ 964.863839] env[62235]: _type = "Task" [ 964.863839] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.876770] env[62235]: DEBUG oslo_vmware.api [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.026470] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3a94c639-baed-4ee4-8a30-d0e29d4b2ba9 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "062bbe32-3edd-4b06-b7d9-882e798b3a7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.478s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.070751] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.174455] env[62235]: INFO nova.compute.manager [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Detaching volume ff64a51b-f62a-4682-83fb-4e26eae16346 [ 965.207174] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc05395-16f1-4f1e-92cd-2c09f6924823 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.210760] env[62235]: INFO nova.virt.block_device [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Attempting to driver detach volume ff64a51b-f62a-4682-83fb-4e26eae16346 from mountpoint /dev/sdb [ 965.211000] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 965.211206] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273467', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'name': 'volume-ff64a51b-f62a-4682-83fb-4e26eae16346', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0e9182f8-81e1-4ca2-97b6-0066faba0b22', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'serial': 'ff64a51b-f62a-4682-83fb-4e26eae16346'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 965.212814] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda959a1-e95c-4417-bf6a-2c02df2f1ca3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.236940] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fd8634-e387-44bb-9d85-9154f4166523 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.241621] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d07a80-1715-4875-a224-766f103e2454 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.276792] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae321ae-b44c-4dab-8f60-cabb69b12db4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.283772] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d47e96e-4d60-4ce1-9e3d-5d99ec448684 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.308706] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981e448e-732c-451e-98a4-4cd8fb198758 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.314262] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234d2d71-5833-41b6-b11e-987cbb8f4158 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.327868] env[62235]: DEBUG nova.compute.provider_tree [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.339395] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] The volume has not been displaced from its original location: [datastore1] volume-ff64a51b-f62a-4682-83fb-4e26eae16346/volume-ff64a51b-f62a-4682-83fb-4e26eae16346.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 965.344417] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 965.345718] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7fbd5e5-e9ae-4b1e-a251-b29c3e5faafd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.365211] env[62235]: DEBUG oslo_vmware.api [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 965.365211] env[62235]: value = "task-1272022" [ 965.365211] env[62235]: _type = "Task" [ 965.365211] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.368419] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-2068403d-5870-4daa-b4d6-acaf39209488-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.368650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-2068403d-5870-4daa-b4d6-acaf39209488-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.379113] env[62235]: DEBUG oslo_vmware.api [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272022, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.382347] env[62235]: DEBUG oslo_vmware.api [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272021, 'name': PowerOffVM_Task, 'duration_secs': 0.253381} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.382617] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.382789] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.383049] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebcdfcc2-e7d3-4d04-b75c-019f1175a974 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.456646] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.456836] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.457284] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleting the datastore file [datastore1] 79dafae1-e389-4e04-af20-577b8aac06a6 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.457616] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4a859e4-eaab-4f44-b8e5-4857c53f0b6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.465695] env[62235]: DEBUG oslo_vmware.api [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 965.465695] env[62235]: value = "task-1272024" [ 965.465695] env[62235]: _type = "Task" [ 965.465695] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.474306] env[62235]: DEBUG oslo_vmware.api [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.699012] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.699406] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.787690] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "6f803e24-2b2b-4a15-8172-29acde2154f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.787908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.788236] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "6f803e24-2b2b-4a15-8172-29acde2154f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.788435] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.788764] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.792299] env[62235]: INFO nova.compute.manager [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Terminating instance [ 965.794221] env[62235]: DEBUG nova.compute.manager [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 965.794418] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.795287] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39cb508-60ee-4676-9326-d4dc9b8702e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.802800] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.803063] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c94f455d-9c6d-4217-8dc2-58a7821d4784 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.847802] env[62235]: DEBUG nova.scheduler.client.report [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.870841] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.870941] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.871110] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleting the datastore file [datastore2] 6f803e24-2b2b-4a15-8172-29acde2154f8 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.871805] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-458f2cd5-a968-4181-b90f-d9f70d638b94 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.874124] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.874331] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.878137] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6650cc6-3cd6-46c3-b4d9-47c39c31b958 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.881919] env[62235]: DEBUG oslo_vmware.api [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272022, 'name': ReconfigVM_Task, 'duration_secs': 0.252998} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.881919] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 965.887643] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-742bbe47-00c4-4795-b273-baa22531b5ce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.897845] env[62235]: DEBUG oslo_vmware.api [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 965.897845] env[62235]: value = "task-1272026" [ 965.897845] env[62235]: _type = "Task" [ 965.897845] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.915839] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40630b0c-b4a3-4420-9dca-6a7a16ed4438 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.920632] env[62235]: DEBUG oslo_vmware.api [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 965.920632] env[62235]: value = "task-1272027" [ 965.920632] env[62235]: _type = "Task" [ 965.920632] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.942219] env[62235]: DEBUG oslo_vmware.api [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272026, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.947920] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Reconfiguring VM to detach interface {{(pid=62235) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 965.949065] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c67a6a9a-013a-4cf9-ad0b-8fff02fa4702 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.966715] env[62235]: DEBUG oslo_vmware.api [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272027, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.974409] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 965.974409] env[62235]: value = "task-1272028" [ 965.974409] env[62235]: _type = "Task" [ 965.974409] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.977823] env[62235]: DEBUG oslo_vmware.api [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161492} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.982951] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.983116] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.983284] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.983462] env[62235]: INFO nova.compute.manager [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 965.983711] env[62235]: DEBUG oslo.service.loopingcall [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.983914] env[62235]: DEBUG nova.compute.manager [-] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.984022] env[62235]: DEBUG nova.network.neutron [-] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.991364] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.202262] env[62235]: INFO nova.compute.manager [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Detaching volume 99abe792-777d-4235-82bc-9e547c16e586 [ 966.238686] env[62235]: INFO nova.virt.block_device [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Attempting to driver detach volume 99abe792-777d-4235-82bc-9e547c16e586 from mountpoint /dev/sdb [ 966.238853] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 966.239099] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273472', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'name': 'volume-99abe792-777d-4235-82bc-9e547c16e586', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15d41c0f-045d-4e1c-88b5-1841f99f2e74', 'attached_at': '', 'detached_at': '', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'serial': '99abe792-777d-4235-82bc-9e547c16e586'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 966.240077] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a7d372-17a4-425e-81c4-e0386aea38b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.270132] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2f04ec-e46a-427b-b1d6-1a821f7723a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.281633] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6fb935-e4fd-42f9-89d5-43ce10461f45 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.304858] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afc7fd2-f41e-4b6c-b58d-a888923a4ad7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.320962] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] The volume has not been displaced from its original location: [datastore1] volume-99abe792-777d-4235-82bc-9e547c16e586/volume-99abe792-777d-4235-82bc-9e547c16e586.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 966.326389] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 966.326797] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d25fb635-f159-4f1f-abe6-ace0c5d79ff6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.350191] env[62235]: DEBUG oslo_vmware.api [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 966.350191] env[62235]: value = "task-1272029" [ 966.350191] env[62235]: _type = "Task" [ 966.350191] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.360279] env[62235]: DEBUG nova.compute.manager [req-e3978f4b-3092-4766-9438-ceed6608fe4b req-8e30fd9f-215d-41fe-b51a-2a3c1c871890 service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Received event network-vif-deleted-e05bb07a-f9f3-48a9-847a-0b12ba578ea8 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.360581] env[62235]: INFO nova.compute.manager [req-e3978f4b-3092-4766-9438-ceed6608fe4b req-8e30fd9f-215d-41fe-b51a-2a3c1c871890 service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Neutron deleted interface e05bb07a-f9f3-48a9-847a-0b12ba578ea8; detaching it from the instance and deleting it from the info cache [ 966.360717] env[62235]: DEBUG nova.network.neutron [req-e3978f4b-3092-4766-9438-ceed6608fe4b req-8e30fd9f-215d-41fe-b51a-2a3c1c871890 service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.365714] env[62235]: DEBUG oslo_vmware.api [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272029, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.423410] env[62235]: DEBUG oslo_vmware.api [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272026, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194046} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.426798] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.427032] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.427230] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.427412] env[62235]: INFO nova.compute.manager [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Took 0.63 seconds to destroy the instance on the hypervisor. [ 966.427654] env[62235]: DEBUG oslo.service.loopingcall [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.427862] env[62235]: DEBUG nova.compute.manager [-] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 966.427957] env[62235]: DEBUG nova.network.neutron [-] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 966.435105] env[62235]: DEBUG oslo_vmware.api [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272027, 'name': ReconfigVM_Task, 'duration_secs': 0.170567} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.435225] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273467', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'name': 'volume-ff64a51b-f62a-4682-83fb-4e26eae16346', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0e9182f8-81e1-4ca2-97b6-0066faba0b22', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff64a51b-f62a-4682-83fb-4e26eae16346', 'serial': 'ff64a51b-f62a-4682-83fb-4e26eae16346'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 966.491999] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.844018] env[62235]: DEBUG nova.network.neutron [-] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.861090] env[62235]: DEBUG oslo_vmware.api [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272029, 'name': ReconfigVM_Task, 'duration_secs': 0.375385} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.862147] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 966.866827] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.379s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.869600] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afef8cfc-c944-4408-a457-39a9c5cf1c70 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.879762] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.088s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.881501] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb1c79c2-8d7c-4773-a355-ac0fdb930674 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.892580] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b16b9b-d141-4c7d-b1b6-14c767e1a3ee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.906214] env[62235]: DEBUG oslo_vmware.api [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 966.906214] env[62235]: value = "task-1272030" [ 966.906214] env[62235]: _type = "Task" [ 966.906214] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.918831] env[62235]: DEBUG oslo_vmware.api [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.929119] env[62235]: DEBUG nova.compute.manager [req-e3978f4b-3092-4766-9438-ceed6608fe4b req-8e30fd9f-215d-41fe-b51a-2a3c1c871890 service nova] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Detach interface failed, port_id=e05bb07a-f9f3-48a9-847a-0b12ba578ea8, reason: Instance 79dafae1-e389-4e04-af20-577b8aac06a6 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 966.978440] env[62235]: DEBUG nova.objects.instance [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'flavor' on Instance uuid 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.994585] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.182503] env[62235]: DEBUG nova.network.neutron [-] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.346998] env[62235]: INFO nova.compute.manager [-] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Took 1.36 seconds to deallocate network for instance. [ 967.386914] env[62235]: INFO nova.compute.claims [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.425956] env[62235]: DEBUG oslo_vmware.api [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272030, 'name': ReconfigVM_Task, 'duration_secs': 0.257506} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.426290] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273472', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'name': 'volume-99abe792-777d-4235-82bc-9e547c16e586', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15d41c0f-045d-4e1c-88b5-1841f99f2e74', 'attached_at': '', 'detached_at': '', 'volume_id': '99abe792-777d-4235-82bc-9e547c16e586', 'serial': '99abe792-777d-4235-82bc-9e547c16e586'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 967.451018] env[62235]: INFO nova.scheduler.client.report [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted allocation for migration b1367deb-b42b-418c-b261-69119a3fb577 [ 967.496806] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.685591] env[62235]: INFO nova.compute.manager [-] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Took 1.26 seconds to deallocate network for instance. [ 967.853383] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.894765] env[62235]: INFO nova.compute.resource_tracker [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating resource usage from migration 58fa497c-b13d-4311-b06e-5c36066eb749 [ 967.919637] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.957077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4d8d0ca2-9ab8-4ed4-9a80-7bdd5111c95f tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.905s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.960890] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.041s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.961390] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.962133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.962133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.965919] env[62235]: INFO nova.compute.manager [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Terminating instance [ 967.968259] env[62235]: DEBUG nova.compute.manager [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 967.968847] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.969546] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55816c81-7d62-4655-a021-eaf9ff540ffd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.977319] env[62235]: DEBUG nova.objects.instance [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'flavor' on Instance uuid 15d41c0f-045d-4e1c-88b5-1841f99f2e74 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.980800] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.981309] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33d983d8-3c79-4ac4-b93c-9d7a865bf48c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.991403] env[62235]: DEBUG oslo_concurrency.lockutils [None req-03684bc5-9d82-44c7-ab91-97769075a5b8 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.322s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.996242] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.997765] env[62235]: DEBUG oslo_vmware.api [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 967.997765] env[62235]: value = "task-1272031" [ 967.997765] env[62235]: _type = "Task" [ 967.997765] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.011029] env[62235]: DEBUG oslo_vmware.api [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.089388] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06561b0-9c4a-4c0a-a932-26b3a0454650 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.098603] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e21fb6-c483-4ea2-b033-fbe4c5ae6cd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.132478] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7877123e-b7df-45f3-b412-1057ce707c12 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.141099] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6c807d-e2f9-4304-b6af-89d0a1f92cc2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.155752] env[62235]: DEBUG nova.compute.provider_tree [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.192197] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.362321] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "8a6962ba-4406-4ac9-8447-b839903781fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.362581] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "8a6962ba-4406-4ac9-8447-b839903781fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.414051] env[62235]: DEBUG nova.compute.manager [req-bc14222d-d6d0-4826-b452-8a8d07c1586e req-d26d18c2-2aef-4992-b923-8cfebc2abed2 service nova] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Received event network-vif-deleted-566501e6-2585-448d-a888-6a640269ed49 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.499474] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.510113] env[62235]: DEBUG oslo_vmware.api [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272031, 'name': PowerOffVM_Task, 'duration_secs': 0.177018} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.510425] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.510595] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.510861] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-998986c8-af1f-4c0d-8862-71b58176b415 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.587486] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.587808] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.588092] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleting the datastore file [datastore2] 05aae40a-37ae-4bf1-91d0-02a7228fcff2 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.588467] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8acc22e1-f2ba-4d5a-bbfd-8abba1409c24 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.596996] env[62235]: DEBUG oslo_vmware.api [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 968.596996] env[62235]: value = "task-1272033" [ 968.596996] env[62235]: _type = "Task" [ 968.596996] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.604948] env[62235]: DEBUG oslo_vmware.api [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272033, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.659197] env[62235]: DEBUG nova.scheduler.client.report [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.864786] env[62235]: DEBUG nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 968.994912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2816c1d1-20b3-4796-8450-81170f14ed05 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.295s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.996148] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.996457] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.996748] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.997021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.997280] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.998877] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.999509] env[62235]: INFO nova.compute.manager [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Terminating instance [ 969.001243] env[62235]: DEBUG nova.compute.manager [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.001506] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.002332] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca10cf06-10ea-4b88-b5b4-d6f12a2c1cc2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.010007] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.010246] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05b80cef-e84f-46db-b6fd-c375b798de86 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.020042] env[62235]: DEBUG oslo_vmware.api [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 969.020042] env[62235]: value = "task-1272034" [ 969.020042] env[62235]: _type = "Task" [ 969.020042] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.029516] env[62235]: DEBUG oslo_vmware.api [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272034, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.107964] env[62235]: DEBUG oslo_vmware.api [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272033, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128228} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.108369] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.108641] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.108868] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.109112] env[62235]: INFO nova.compute.manager [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Took 1.14 seconds to destroy the instance on the hypervisor. [ 969.109408] env[62235]: DEBUG oslo.service.loopingcall [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.109628] env[62235]: DEBUG nova.compute.manager [-] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.109723] env[62235]: DEBUG nova.network.neutron [-] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 969.164420] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.284s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.164707] env[62235]: INFO nova.compute.manager [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Migrating [ 969.173995] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.103s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.175501] env[62235]: INFO nova.compute.claims [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.387480] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.499715] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.530933] env[62235]: DEBUG oslo_vmware.api [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272034, 'name': PowerOffVM_Task, 'duration_secs': 0.180292} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.531930] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.531930] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.532106] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da5be529-68f2-4ff6-ac5b-2b3f1d608460 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.602933] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.602933] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.602933] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleting the datastore file [datastore1] 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.602933] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67aadcb9-5bd6-4cb5-b02c-e4fd54e84cde {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.611834] env[62235]: DEBUG oslo_vmware.api [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 969.611834] env[62235]: value = "task-1272036" [ 969.611834] env[62235]: _type = "Task" [ 969.611834] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.619121] env[62235]: DEBUG oslo_vmware.api [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272036, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.687877] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.688152] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.688298] env[62235]: DEBUG nova.network.neutron [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.855414] env[62235]: DEBUG nova.network.neutron [-] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.998111] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.998480] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.998577] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.998761] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.998931] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.000561] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.001030] env[62235]: INFO nova.compute.manager [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Terminating instance [ 970.002947] env[62235]: DEBUG nova.compute.manager [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.003171] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.003907] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08baf38-a942-4d72-9c78-c9e0a8eec44d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.010988] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.011228] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e395911a-6713-4034-bc40-6a30cf3262f9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.018822] env[62235]: DEBUG oslo_vmware.api [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 970.018822] env[62235]: value = "task-1272037" [ 970.018822] env[62235]: _type = "Task" [ 970.018822] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.028411] env[62235]: DEBUG oslo_vmware.api [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.121942] env[62235]: DEBUG oslo_vmware.api [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272036, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133571} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.122247] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.122442] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.122647] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.122844] env[62235]: INFO nova.compute.manager [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Took 1.12 seconds to destroy the instance on the hypervisor. [ 970.123138] env[62235]: DEBUG oslo.service.loopingcall [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.123354] env[62235]: DEBUG nova.compute.manager [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.123447] env[62235]: DEBUG nova.network.neutron [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.358384] env[62235]: INFO nova.compute.manager [-] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Took 1.25 seconds to deallocate network for instance. [ 970.394549] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcba9c83-2fcf-4b3c-8819-17a359d8c915 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.403007] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87142316-febe-445a-9673-3f8c56d132b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.441716] env[62235]: DEBUG nova.network.neutron [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.443439] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7898a131-9ba1-4aac-b36c-121fecbecf84 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.448121] env[62235]: DEBUG nova.compute.manager [req-a13054ec-246a-48fa-9a97-1e4e7f1c6132 req-19eb8854-28c1-448b-9ccb-21a803722562 service nova] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Received event network-vif-deleted-e3865625-b03d-42d2-8ba3-8829b7996727 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.455726] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f237d88a-d0b8-4e8b-a258-d64db111f57e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.473121] env[62235]: DEBUG nova.compute.provider_tree [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.498461] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.529317] env[62235]: DEBUG oslo_vmware.api [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272037, 'name': PowerOffVM_Task, 'duration_secs': 0.198449} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.529539] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.529706] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.529968] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2a1fd62-0045-4ea5-b8e7-b108569d6249 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.594969] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.595383] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.595597] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleting the datastore file [datastore2] 15d41c0f-045d-4e1c-88b5-1841f99f2e74 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.596101] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-174a28d2-0295-4ed9-9213-37a6b35bf4e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.605910] env[62235]: DEBUG oslo_vmware.api [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 970.605910] env[62235]: value = "task-1272039" [ 970.605910] env[62235]: _type = "Task" [ 970.605910] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.619458] env[62235]: DEBUG oslo_vmware.api [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272039, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.866536] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.948656] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.976470] env[62235]: DEBUG nova.scheduler.client.report [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.998689] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.119608] env[62235]: DEBUG oslo_vmware.api [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272039, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149681} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.121502] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.121866] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.122215] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.122580] env[62235]: INFO nova.compute.manager [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Took 1.12 seconds to destroy the instance on the hypervisor. [ 971.123089] env[62235]: DEBUG oslo.service.loopingcall [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.123412] env[62235]: DEBUG nova.compute.manager [-] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.124161] env[62235]: DEBUG nova.network.neutron [-] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 971.321379] env[62235]: DEBUG nova.network.neutron [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.484023] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.484023] env[62235]: DEBUG nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 971.484940] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.632s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.485316] env[62235]: DEBUG nova.objects.instance [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lazy-loading 'resources' on Instance uuid 79dafae1-e389-4e04-af20-577b8aac06a6 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.500612] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.824183] env[62235]: INFO nova.compute.manager [-] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Took 1.70 seconds to deallocate network for instance. [ 971.987994] env[62235]: DEBUG nova.compute.utils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 971.993435] env[62235]: DEBUG nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 971.993435] env[62235]: DEBUG nova.network.neutron [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 972.008259] env[62235]: DEBUG oslo_vmware.api [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272028, 'name': ReconfigVM_Task, 'duration_secs': 5.834389} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.008836] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.009061] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Reconfigured VM to detach interface {{(pid=62235) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 972.055798] env[62235]: DEBUG nova.policy [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11fe623e504d4729a149cf666aed4c67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b690e5ef3dac4a07879b4b7fe2249e89', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 972.067756] env[62235]: DEBUG nova.network.neutron [-] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.205608] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445d92e8-ddc1-445c-b5ab-813548eff168 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.218651] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3b72e5-9546-4764-8f9b-aaf48665577d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.269923] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca8e587-befe-4a62-b45e-d94e8c4e69b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.281318] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9607cc38-06a5-478c-8857-3f6284de0a6f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.296837] env[62235]: DEBUG nova.compute.provider_tree [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.330612] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.352363] env[62235]: DEBUG nova.network.neutron [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Successfully created port: 3e813a1d-b62c-400a-a376-ec445a16c4ca {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 972.465976] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bb85a0-d29d-4208-956f-5c4631ec91c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.485091] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance '3ea7db36-ebde-4523-b803-ad121a00269b' progress to 0 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.490608] env[62235]: DEBUG nova.compute.manager [req-4d0263b9-2111-4c2e-adda-00c42c0ca82f req-408a6c7b-ccc3-4ac9-8f15-c88b324eab90 service nova] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Received event network-vif-deleted-f6a488c2-b3b0-44f9-8f39-c169dd5f2dd4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.490809] env[62235]: DEBUG nova.compute.manager [req-4d0263b9-2111-4c2e-adda-00c42c0ca82f req-408a6c7b-ccc3-4ac9-8f15-c88b324eab90 service nova] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Received event network-vif-deleted-9af619cc-4e60-4097-96e6-b957861ba812 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.492772] env[62235]: DEBUG nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 972.572925] env[62235]: INFO nova.compute.manager [-] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Took 1.45 seconds to deallocate network for instance. [ 972.862335] env[62235]: ERROR nova.scheduler.client.report [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [req-1b38285a-bea7-457e-81f6-882009a4ca97] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1b38285a-bea7-457e-81f6-882009a4ca97"}]} [ 972.862335] env[62235]: DEBUG nova.scheduler.client.report [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 972.866502] env[62235]: DEBUG nova.scheduler.client.report [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 972.866722] env[62235]: DEBUG nova.compute.provider_tree [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.882200] env[62235]: DEBUG nova.scheduler.client.report [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 972.902271] env[62235]: DEBUG nova.scheduler.client.report [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 972.994191] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.995405] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70e285dd-a39e-4014-95fc-dfc87854dc6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.007567] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 973.007567] env[62235]: value = "task-1272040" [ 973.007567] env[62235]: _type = "Task" [ 973.007567] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.018036] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272040, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.079804] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.094188] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b293edc-f537-491a-acf3-858f25245d38 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.103416] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c222f769-200f-4580-aa5e-134d7891bed6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.136869] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753f5707-45d6-460a-9fed-fbd444f21b77 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.145267] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccc965d-aa36-4a8e-bc5b-e3ed89e43c3d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.160860] env[62235]: DEBUG nova.compute.provider_tree [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.343869] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.343869] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.343869] env[62235]: DEBUG nova.network.neutron [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.502239] env[62235]: DEBUG nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 973.518299] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272040, 'name': PowerOffVM_Task, 'duration_secs': 0.235088} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.518599] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.518797] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance '3ea7db36-ebde-4523-b803-ad121a00269b' progress to 17 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 973.531663] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 973.531903] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 973.532085] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.532381] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 973.532563] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.532718] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 973.532931] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 973.533132] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 973.533309] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 973.533471] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 973.533645] env[62235]: DEBUG nova.virt.hardware [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 973.534493] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b956feba-c4ca-4def-a09c-30d170e55a04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.544547] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ded01b-532a-41f0-83ad-018cb3a4dcff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.695596] env[62235]: DEBUG nova.scheduler.client.report [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Updated inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 973.695879] env[62235]: DEBUG nova.compute.provider_tree [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 112 to 113 during operation: update_inventory {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 973.696076] env[62235]: DEBUG nova.compute.provider_tree [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.823758] env[62235]: DEBUG nova.network.neutron [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Successfully updated port: 3e813a1d-b62c-400a-a376-ec445a16c4ca {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 974.027128] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.027485] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.027594] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.027795] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.027965] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.031479] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.031596] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.031741] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.031926] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.032267] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.032349] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.040415] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afd90dd9-c787-4825-b0b4-9acad6378b3b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.053206] env[62235]: DEBUG nova.compute.manager [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-changed-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.053406] env[62235]: DEBUG nova.compute.manager [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing instance network info cache due to event network-changed-64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 974.053601] env[62235]: DEBUG oslo_concurrency.lockutils [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] Acquiring lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.061811] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 974.061811] env[62235]: value = "task-1272041" [ 974.061811] env[62235]: _type = "Task" [ 974.061811] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.071820] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272041, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.088693] env[62235]: INFO nova.network.neutron [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Port ac6a259e-ebe5-4b85-b07f-38fe6b16c73b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 974.089145] env[62235]: DEBUG nova.network.neutron [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.186095] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 974.186405] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 974.200726] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.716s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.204262] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.011s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.204482] env[62235]: DEBUG nova.objects.instance [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'resources' on Instance uuid 6f803e24-2b2b-4a15-8172-29acde2154f8 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.229136] env[62235]: INFO nova.scheduler.client.report [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleted allocations for instance 79dafae1-e389-4e04-af20-577b8aac06a6 [ 974.326057] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.326309] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.326383] env[62235]: DEBUG nova.network.neutron [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 974.571022] env[62235]: DEBUG nova.compute.manager [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Received event network-vif-plugged-3e813a1d-b62c-400a-a376-ec445a16c4ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.571195] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] Acquiring lock "550945fe-52be-408d-9f42-f5cd5e9c489f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.571400] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.571527] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.571710] env[62235]: DEBUG nova.compute.manager [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] No waiting events found dispatching network-vif-plugged-3e813a1d-b62c-400a-a376-ec445a16c4ca {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 974.571882] env[62235]: WARNING nova.compute.manager [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Received unexpected event network-vif-plugged-3e813a1d-b62c-400a-a376-ec445a16c4ca for instance with vm_state building and task_state spawning. [ 974.572061] env[62235]: DEBUG nova.compute.manager [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Received event network-changed-3e813a1d-b62c-400a-a376-ec445a16c4ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.572287] env[62235]: DEBUG nova.compute.manager [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Refreshing instance network info cache due to event network-changed-3e813a1d-b62c-400a-a376-ec445a16c4ca. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 974.572467] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] Acquiring lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.576071] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272041, 'name': ReconfigVM_Task, 'duration_secs': 0.204206} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.576361] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance '3ea7db36-ebde-4523-b803-ad121a00269b' progress to 33 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.591483] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.593898] env[62235]: DEBUG oslo_concurrency.lockutils [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] Acquired lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.593898] env[62235]: DEBUG nova.network.neutron [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Refreshing network info cache for port 64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 974.693991] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 974.694175] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 974.741267] env[62235]: DEBUG oslo_concurrency.lockutils [None req-20f13cbc-0314-4d8d-bfac-0f74dd211d05 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "79dafae1-e389-4e04-af20-577b8aac06a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.909s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.791037] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-a3336b20-5737-4e5a-94a1-19fec140aa60-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.791299] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-a3336b20-5737-4e5a-94a1-19fec140aa60-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.791657] env[62235]: DEBUG nova.objects.instance [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'flavor' on Instance uuid a3336b20-5737-4e5a-94a1-19fec140aa60 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.856901] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e160dbc-d48d-4f08-9875-4c50c72fc14e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.860165] env[62235]: DEBUG nova.network.neutron [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 974.867325] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e61cfa5-30f9-4c04-a185-138c84cb53b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.899310] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4239a7-062b-4c9b-8848-ca8482f94e73 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.907257] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2284d344-8818-42f4-9f10-bac552036334 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.922321] env[62235]: DEBUG nova.compute.provider_tree [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.020561] env[62235]: DEBUG nova.network.neutron [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Updating instance_info_cache with network_info: [{"id": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "address": "fa:16:3e:ec:f5:f9", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e813a1d-b6", "ovs_interfaceid": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.082283] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.082624] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.082689] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.082868] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.083029] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.083207] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.083430] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.083605] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.083778] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.083943] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.084134] env[62235]: DEBUG nova.virt.hardware [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.089711] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Reconfiguring VM instance instance-0000003a to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 975.090010] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed8dbc80-2f96-4af3-b012-285365436de1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.103576] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1b5f3d98-66c7-46f2-8e92-3d0211f0935c tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-2068403d-5870-4daa-b4d6-acaf39209488-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.735s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.113795] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 975.113795] env[62235]: value = "task-1272042" [ 975.113795] env[62235]: _type = "Task" [ 975.113795] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.124963] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.310232] env[62235]: DEBUG nova.network.neutron [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updated VIF entry in instance network info cache for port 64d54817-98a8-4e2e-b317-957980c86633. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 975.310626] env[62235]: DEBUG nova.network.neutron [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [{"id": "64d54817-98a8-4e2e-b317-957980c86633", "address": "fa:16:3e:01:0b:89", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64d54817-98", "ovs_interfaceid": "64d54817-98a8-4e2e-b317-957980c86633", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.426571] env[62235]: DEBUG nova.scheduler.client.report [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 975.433198] env[62235]: DEBUG nova.objects.instance [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'pci_requests' on Instance uuid a3336b20-5737-4e5a-94a1-19fec140aa60 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.523522] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.524037] env[62235]: DEBUG nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Instance network_info: |[{"id": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "address": "fa:16:3e:ec:f5:f9", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e813a1d-b6", "ovs_interfaceid": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 975.524185] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] Acquired lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.524344] env[62235]: DEBUG nova.network.neutron [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Refreshing network info cache for port 3e813a1d-b62c-400a-a376-ec445a16c4ca {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 975.526013] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:f5:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e813a1d-b62c-400a-a376-ec445a16c4ca', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.533881] env[62235]: DEBUG oslo.service.loopingcall [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.537214] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 975.537997] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ef661ce-9c88-4100-880a-cf28e7ef6a7f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.559938] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 975.559938] env[62235]: value = "task-1272043" [ 975.559938] env[62235]: _type = "Task" [ 975.559938] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.568376] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272043, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.623971] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272042, 'name': ReconfigVM_Task, 'duration_secs': 0.23444} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.624278] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Reconfigured VM instance instance-0000003a to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 975.625060] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d95130-605f-4ac7-9978-9db54d8faf0b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.649473] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 3ea7db36-ebde-4523-b803-ad121a00269b/3ea7db36-ebde-4523-b803-ad121a00269b.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.652257] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2640cca7-bf48-49d0-aaa5-6c5af8d3a68a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.672025] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 975.672025] env[62235]: value = "task-1272044" [ 975.672025] env[62235]: _type = "Task" [ 975.672025] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.687707] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272044, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.766022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.766311] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.766534] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.766722] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.766963] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.771057] env[62235]: INFO nova.compute.manager [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Terminating instance [ 975.772994] env[62235]: DEBUG nova.compute.manager [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.773238] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 975.774155] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29ed51b-fb97-47e0-b69e-d64798112693 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.782672] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.782947] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0942a20-b34f-460f-a9cf-c4dc72a15e37 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.791246] env[62235]: DEBUG oslo_vmware.api [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 975.791246] env[62235]: value = "task-1272045" [ 975.791246] env[62235]: _type = "Task" [ 975.791246] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.799914] env[62235]: DEBUG oslo_vmware.api [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.808892] env[62235]: DEBUG nova.network.neutron [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Updated VIF entry in instance network info cache for port 3e813a1d-b62c-400a-a376-ec445a16c4ca. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 975.809281] env[62235]: DEBUG nova.network.neutron [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Updating instance_info_cache with network_info: [{"id": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "address": "fa:16:3e:ec:f5:f9", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e813a1d-b6", "ovs_interfaceid": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.813571] env[62235]: DEBUG oslo_concurrency.lockutils [req-18d6b85c-e25c-4785-9add-32ebf1f2a7ac req-058a505f-3588-4db9-82d7-1105b1974555 service nova] Releasing lock "refresh_cache-2068403d-5870-4daa-b4d6-acaf39209488" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.931340] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.728s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.934033] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.547s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.935479] env[62235]: INFO nova.compute.claims [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.938233] env[62235]: DEBUG nova.objects.base [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 975.938430] env[62235]: DEBUG nova.network.neutron [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 975.987486] env[62235]: INFO nova.scheduler.client.report [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted allocations for instance 6f803e24-2b2b-4a15-8172-29acde2154f8 [ 976.022398] env[62235]: DEBUG nova.policy [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a017d5f37c444d6aa59ab31d43e6215', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e885dc2e5404662a081bb3855919b0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 976.071176] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272043, 'name': CreateVM_Task, 'duration_secs': 0.389343} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.071392] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 976.072088] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.072294] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.072628] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 976.073842] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b73690e5-36fa-4c87-ae21-2ff499b9126e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.076421] env[62235]: DEBUG nova.compute.manager [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.076539] env[62235]: DEBUG nova.compute.manager [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing instance network info cache due to event network-changed-1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.076742] env[62235]: DEBUG oslo_concurrency.lockutils [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.076888] env[62235]: DEBUG oslo_concurrency.lockutils [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.077083] env[62235]: DEBUG nova.network.neutron [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 976.082461] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 976.082461] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210fb4f-259e-f7a0-df26-88f48e88156c" [ 976.082461] env[62235]: _type = "Task" [ 976.082461] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.092015] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210fb4f-259e-f7a0-df26-88f48e88156c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.182527] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272044, 'name': ReconfigVM_Task, 'duration_secs': 0.360896} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.182841] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 3ea7db36-ebde-4523-b803-ad121a00269b/3ea7db36-ebde-4523-b803-ad121a00269b.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.183122] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance '3ea7db36-ebde-4523-b803-ad121a00269b' progress to 50 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 976.301599] env[62235]: DEBUG oslo_vmware.api [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272045, 'name': PowerOffVM_Task, 'duration_secs': 0.272384} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.301887] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 976.302072] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 976.302449] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1996c858-dbb5-4972-b655-9082a60370a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.311724] env[62235]: DEBUG oslo_concurrency.lockutils [req-ba00dcf4-7a9d-40f7-8b53-ffc44a0bc3e0 req-a64aa31f-8774-4a88-99d1-467d36d800ab service nova] Releasing lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.372828] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 976.373077] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 976.373364] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleting the datastore file [datastore1] 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.373689] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-719e887e-5137-4c43-bcfc-f256e01fc19f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.380518] env[62235]: DEBUG oslo_vmware.api [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 976.380518] env[62235]: value = "task-1272047" [ 976.380518] env[62235]: _type = "Task" [ 976.380518] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.390596] env[62235]: DEBUG oslo_vmware.api [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272047, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.496549] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d032a959-7570-4048-b69a-b1a3b1d33a8f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "6f803e24-2b2b-4a15-8172-29acde2154f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.709s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.593675] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5210fb4f-259e-f7a0-df26-88f48e88156c, 'name': SearchDatastore_Task, 'duration_secs': 0.010258} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.594201] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.594634] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.595063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.595364] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.595713] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.596594] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6fcd3ba0-6235-43c7-b290-73e4a40c5639 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.607806] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.608025] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.608996] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce430fcd-b94a-4e0a-bde2-a50473a1589d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.615624] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 976.615624] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5236ef17-bd32-27a2-273d-7d000c2b7293" [ 976.615624] env[62235]: _type = "Task" [ 976.615624] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.626377] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5236ef17-bd32-27a2-273d-7d000c2b7293, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.690888] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e2297b-f40d-4c5b-b160-f764ca097bce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.714849] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdcf182f-08eb-486f-a4db-fda82fc5e84e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.735014] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance '3ea7db36-ebde-4523-b803-ad121a00269b' progress to 67 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 976.811971] env[62235]: DEBUG nova.network.neutron [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updated VIF entry in instance network info cache for port 1bc70ab7-baf7-4b1d-b945-41463b6ea517. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 976.812493] env[62235]: DEBUG nova.network.neutron [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.891808] env[62235]: DEBUG oslo_vmware.api [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272047, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13806} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.891808] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.891808] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.891808] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.892143] env[62235]: INFO nova.compute.manager [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 976.892230] env[62235]: DEBUG oslo.service.loopingcall [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.892427] env[62235]: DEBUG nova.compute.manager [-] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.892524] env[62235]: DEBUG nova.network.neutron [-] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 977.130843] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5236ef17-bd32-27a2-273d-7d000c2b7293, 'name': SearchDatastore_Task, 'duration_secs': 0.010769} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.136096] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05e03f31-9021-410e-9e9a-cf6c9bd72a2b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.141506] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.141824] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.146105] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768a278c-cc77-4d74-80a6-8b37e0eaae6d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.152893] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 977.152893] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527270b3-3829-02df-5d7e-89f215b341ae" [ 977.152893] env[62235]: _type = "Task" [ 977.152893] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.160034] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0405c811-5444-4d5a-a991-cebb48a49e26 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.167890] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527270b3-3829-02df-5d7e-89f215b341ae, 'name': SearchDatastore_Task, 'duration_secs': 0.011075} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.168233] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.168465] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/550945fe-52be-408d-9f42-f5cd5e9c489f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.168740] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b35c455-3f5f-4886-83bf-475052d9f64c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.197890] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76bae0aa-ac3c-415b-ae47-1c59311121c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.202180] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 977.202180] env[62235]: value = "task-1272048" [ 977.202180] env[62235]: _type = "Task" [ 977.202180] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.209643] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94479bc9-3dab-4d7a-a693-597fb0aaf557 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.216971] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272048, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.230152] env[62235]: DEBUG nova.compute.provider_tree [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.233790] env[62235]: DEBUG nova.compute.manager [req-f2d028f6-8754-49ca-a48d-765d46750618 req-55feee6c-4bb7-4a06-ad24-c389d81d5c6c service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Received event network-vif-deleted-21adaaf5-f619-4e50-9341-0cb69b5c91c6 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.233790] env[62235]: INFO nova.compute.manager [req-f2d028f6-8754-49ca-a48d-765d46750618 req-55feee6c-4bb7-4a06-ad24-c389d81d5c6c service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Neutron deleted interface 21adaaf5-f619-4e50-9341-0cb69b5c91c6; detaching it from the instance and deleting it from the info cache [ 977.233790] env[62235]: DEBUG nova.network.neutron [req-f2d028f6-8754-49ca-a48d-765d46750618 req-55feee6c-4bb7-4a06-ad24-c389d81d5c6c service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.280998] env[62235]: DEBUG nova.network.neutron [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Port eed03567-8c20-4e56-902c-fcd6bb25b56b binding to destination host cpu-1 is already ACTIVE {{(pid=62235) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 977.314726] env[62235]: DEBUG oslo_concurrency.lockutils [req-a26c8fef-3ffa-424d-a935-4126b75fa7c4 req-0e601e93-040f-4ca8-bf05-b749da37816a service nova] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.581044] env[62235]: DEBUG nova.network.neutron [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Successfully updated port: ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.644786] env[62235]: DEBUG nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.703154] env[62235]: DEBUG nova.network.neutron [-] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.715730] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272048, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.735832] env[62235]: DEBUG nova.scheduler.client.report [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.740028] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d66880b-59cb-4764-8452-495e19a627b4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.753605] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31980fe-65d6-466b-8d00-ce550080e3e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.789368] env[62235]: DEBUG nova.compute.manager [req-f2d028f6-8754-49ca-a48d-765d46750618 req-55feee6c-4bb7-4a06-ad24-c389d81d5c6c service nova] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Detach interface failed, port_id=21adaaf5-f619-4e50-9341-0cb69b5c91c6, reason: Instance 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 978.083857] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.083857] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.084013] env[62235]: DEBUG nova.network.neutron [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.104745] env[62235]: DEBUG nova.compute.manager [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-vif-plugged-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.104969] env[62235]: DEBUG oslo_concurrency.lockutils [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] Acquiring lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.105341] env[62235]: DEBUG oslo_concurrency.lockutils [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.105458] env[62235]: DEBUG oslo_concurrency.lockutils [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.105805] env[62235]: DEBUG nova.compute.manager [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] No waiting events found dispatching network-vif-plugged-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.105870] env[62235]: WARNING nova.compute.manager [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received unexpected event network-vif-plugged-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b for instance with vm_state active and task_state None. [ 978.106058] env[62235]: DEBUG nova.compute.manager [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-changed-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.106200] env[62235]: DEBUG nova.compute.manager [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing instance network info cache due to event network-changed-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.106372] env[62235]: DEBUG oslo_concurrency.lockutils [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.165328] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.211069] env[62235]: INFO nova.compute.manager [-] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Took 1.32 seconds to deallocate network for instance. [ 978.216646] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272048, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66772} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.216646] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/550945fe-52be-408d-9f42-f5cd5e9c489f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.216825] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.217048] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-519cd0fe-2694-4b3d-ac09-d7450d2e56fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.226910] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 978.226910] env[62235]: value = "task-1272049" [ 978.226910] env[62235]: _type = "Task" [ 978.226910] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.235688] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272049, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.244469] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.244778] env[62235]: DEBUG nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.247490] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.381s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.247694] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.249778] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.919s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.250203] env[62235]: DEBUG nova.objects.instance [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'resources' on Instance uuid 0e9182f8-81e1-4ca2-97b6-0066faba0b22 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.271451] env[62235]: INFO nova.scheduler.client.report [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted allocations for instance 05aae40a-37ae-4bf1-91d0-02a7228fcff2 [ 978.303621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.303861] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.304063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.622649] env[62235]: WARNING nova.network.neutron [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] 695860c9-9102-4053-ad1c-75a1c4ac4b12 already exists in list: networks containing: ['695860c9-9102-4053-ad1c-75a1c4ac4b12']. ignoring it [ 978.722926] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.737498] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272049, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071101} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.737806] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.738590] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9c60b1-97bd-486a-ba93-bbd9ed64d8ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.755950] env[62235]: DEBUG nova.compute.utils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.766421] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/550945fe-52be-408d-9f42-f5cd5e9c489f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.767623] env[62235]: DEBUG nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.772656] env[62235]: DEBUG nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.772656] env[62235]: DEBUG nova.network.neutron [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 978.773973] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-729fe6cf-9b81-4ba6-83ae-a01e2fe72c63 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.793800] env[62235]: DEBUG oslo_concurrency.lockutils [None req-670689da-14fb-44c6-b734-999f29d44671 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "05aae40a-37ae-4bf1-91d0-02a7228fcff2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.833s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.799947] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 978.799947] env[62235]: value = "task-1272050" [ 978.799947] env[62235]: _type = "Task" [ 978.799947] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.812600] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.838032] env[62235]: DEBUG nova.policy [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b64f6d6696c042079cced4c83d161f6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4469550d32b4482a9d5e9244cbd681e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.945507] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471788d8-9cfc-4a9d-bd26-300f25e0f9d7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.955204] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e302b4a6-c183-40f8-836c-a40a0ae5ba56 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.996573] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492994a9-e673-456e-a655-0893ea0d4618 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.000951] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5614a71-81b0-4e2d-9b0c-a7fba29a0046 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.017091] env[62235]: DEBUG nova.compute.provider_tree [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.091558] env[62235]: DEBUG nova.network.neutron [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Successfully created port: 511a3ca7-0c6b-4e8f-a157-143ce99d2382 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.120852] env[62235]: DEBUG nova.network.neutron [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "address": "fa:16:3e:77:07:e1", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac6a259e-eb", "ovs_interfaceid": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.240727] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.240727] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.240727] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 979.313827] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272050, 'name': ReconfigVM_Task, 'duration_secs': 0.294139} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.317103] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/550945fe-52be-408d-9f42-f5cd5e9c489f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.318456] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ace618cf-b508-49eb-896a-2fed9542c59f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.327118] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 979.327118] env[62235]: value = "task-1272051" [ 979.327118] env[62235]: _type = "Task" [ 979.327118] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.337857] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272051, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.359795] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.360098] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.360517] env[62235]: DEBUG nova.network.neutron [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.536415] env[62235]: ERROR nova.scheduler.client.report [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [req-5057a86d-d8c2-4965-aaf5-2212d4c73b45] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5057a86d-d8c2-4965-aaf5-2212d4c73b45"}]} [ 979.553350] env[62235]: DEBUG nova.scheduler.client.report [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 979.567035] env[62235]: DEBUG nova.scheduler.client.report [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 979.567369] env[62235]: DEBUG nova.compute.provider_tree [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.578456] env[62235]: DEBUG nova.scheduler.client.report [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 979.595237] env[62235]: DEBUG nova.scheduler.client.report [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 979.625870] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.626999] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.627212] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.627695] env[62235]: DEBUG oslo_concurrency.lockutils [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.627893] env[62235]: DEBUG nova.network.neutron [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Refreshing network info cache for port ac6a259e-ebe5-4b85-b07f-38fe6b16c73b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.629537] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8850fa08-13c6-42e2-af6c-cc5f486f1480 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.649269] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.649540] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.649708] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.649895] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.650073] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.650282] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.650511] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.650682] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.650854] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.651034] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.651221] env[62235]: DEBUG nova.virt.hardware [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.657397] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Reconfiguring VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 979.660690] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de0fc8b8-a9cc-4cfe-bbf6-d099bd63148c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.679408] env[62235]: DEBUG oslo_vmware.api [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 979.679408] env[62235]: value = "task-1272052" [ 979.679408] env[62235]: _type = "Task" [ 979.679408] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.690486] env[62235]: DEBUG oslo_vmware.api [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272052, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.770276] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 979.789837] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfd7d0f-7c2c-42eb-9038-e6a5faddfc93 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.795553] env[62235]: DEBUG nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.800242] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad4aee6-865f-41cb-9951-5a7afe2b1f26 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.838500] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243efd81-cefa-47c8-a321-1f911c1b12f9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.847802] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.848092] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.848220] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.848424] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.848644] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.848752] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.848950] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.849086] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.849259] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.849431] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.849604] env[62235]: DEBUG nova.virt.hardware [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.853642] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a47d16-af9b-4488-962c-3e765993337c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.857803] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a296697-b8ec-4a3a-a0b6-5a6e51059550 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.860461] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272051, 'name': Rename_Task, 'duration_secs': 0.182467} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.860747] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.861564] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d4ffa528-a804-4578-9963-5bb2a5848f1f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.875858] env[62235]: DEBUG nova.compute.provider_tree [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.880031] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24cae64-ac6b-4f3c-abfb-6393f02e5b5e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.885926] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 979.885926] env[62235]: value = "task-1272053" [ 979.885926] env[62235]: _type = "Task" [ 979.885926] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.904363] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272053, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.142710] env[62235]: DEBUG nova.network.neutron [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.190662] env[62235]: DEBUG oslo_vmware.api [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272052, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.403221] env[62235]: DEBUG oslo_vmware.api [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272053, 'name': PowerOnVM_Task, 'duration_secs': 0.445159} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.403604] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.403885] env[62235]: INFO nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Took 6.90 seconds to spawn the instance on the hypervisor. [ 980.404162] env[62235]: DEBUG nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.405284] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7dcc23-4030-4b32-8b6e-4b1f1be11915 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.419165] env[62235]: DEBUG nova.scheduler.client.report [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updated inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with generation 114 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 980.419512] env[62235]: DEBUG nova.compute.provider_tree [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 114 to 115 during operation: update_inventory {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 980.419768] env[62235]: DEBUG nova.compute.provider_tree [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.480232] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.517391] env[62235]: DEBUG nova.network.neutron [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updated VIF entry in instance network info cache for port ac6a259e-ebe5-4b85-b07f-38fe6b16c73b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.517952] env[62235]: DEBUG nova.network.neutron [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "address": "fa:16:3e:77:07:e1", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac6a259e-eb", "ovs_interfaceid": "ac6a259e-ebe5-4b85-b07f-38fe6b16c73b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.571195] env[62235]: DEBUG nova.network.neutron [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Successfully updated port: 511a3ca7-0c6b-4e8f-a157-143ce99d2382 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.611110] env[62235]: DEBUG nova.compute.manager [req-1e54e728-bdea-4d9c-93fb-6502ef885c3a req-63570540-e51e-4df0-a7de-7044d8eaef48 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Received event network-vif-plugged-511a3ca7-0c6b-4e8f-a157-143ce99d2382 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.611410] env[62235]: DEBUG oslo_concurrency.lockutils [req-1e54e728-bdea-4d9c-93fb-6502ef885c3a req-63570540-e51e-4df0-a7de-7044d8eaef48 service nova] Acquiring lock "8a6962ba-4406-4ac9-8447-b839903781fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.611626] env[62235]: DEBUG oslo_concurrency.lockutils [req-1e54e728-bdea-4d9c-93fb-6502ef885c3a req-63570540-e51e-4df0-a7de-7044d8eaef48 service nova] Lock "8a6962ba-4406-4ac9-8447-b839903781fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.611761] env[62235]: DEBUG oslo_concurrency.lockutils [req-1e54e728-bdea-4d9c-93fb-6502ef885c3a req-63570540-e51e-4df0-a7de-7044d8eaef48 service nova] Lock "8a6962ba-4406-4ac9-8447-b839903781fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.611935] env[62235]: DEBUG nova.compute.manager [req-1e54e728-bdea-4d9c-93fb-6502ef885c3a req-63570540-e51e-4df0-a7de-7044d8eaef48 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] No waiting events found dispatching network-vif-plugged-511a3ca7-0c6b-4e8f-a157-143ce99d2382 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.612136] env[62235]: WARNING nova.compute.manager [req-1e54e728-bdea-4d9c-93fb-6502ef885c3a req-63570540-e51e-4df0-a7de-7044d8eaef48 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Received unexpected event network-vif-plugged-511a3ca7-0c6b-4e8f-a157-143ce99d2382 for instance with vm_state building and task_state spawning. [ 980.645058] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.690987] env[62235]: DEBUG oslo_vmware.api [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272052, 'name': ReconfigVM_Task, 'duration_secs': 0.612801} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.691514] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.691740] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Reconfigured VM to attach interface {{(pid=62235) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 980.924757] env[62235]: INFO nova.compute.manager [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Took 15.87 seconds to build instance. [ 980.926389] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.676s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.928573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.849s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.928728] env[62235]: DEBUG nova.objects.instance [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'resources' on Instance uuid 15d41c0f-045d-4e1c-88b5-1841f99f2e74 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.946010] env[62235]: INFO nova.scheduler.client.report [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted allocations for instance 0e9182f8-81e1-4ca2-97b6-0066faba0b22 [ 980.983134] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-15d41c0f-045d-4e1c-88b5-1841f99f2e74" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.983905] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 980.983905] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.983905] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.984183] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.984183] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.984280] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.984466] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.984702] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 980.984872] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 981.020177] env[62235]: DEBUG oslo_concurrency.lockutils [req-fdf9c125-6c96-4a64-9f3a-8b1181824fec req-252410cb-d2e1-4e62-9425-9b9b028d337a service nova] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.074050] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "refresh_cache-8a6962ba-4406-4ac9-8447-b839903781fe" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.074207] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "refresh_cache-8a6962ba-4406-4ac9-8447-b839903781fe" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.074255] env[62235]: DEBUG nova.network.neutron [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.167462] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30bf262-fa55-4d0c-8ed7-dca7ac9233ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.187893] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e639b526-5fbd-4747-bf8c-a687630cdbe3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.196255] env[62235]: DEBUG oslo_concurrency.lockutils [None req-3b0f4462-fab8-4057-8c4b-cd8840bd6895 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-a3336b20-5737-4e5a-94a1-19fec140aa60-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.405s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.197952] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance '3ea7db36-ebde-4523-b803-ad121a00269b' progress to 83 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 981.429465] env[62235]: DEBUG oslo_concurrency.lockutils [None req-904b920a-96d4-45c4-aaa4-70942d41193c tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.388s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.454934] env[62235]: DEBUG oslo_concurrency.lockutils [None req-56fb8759-921b-43de-b63c-fd02a9eff741 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "0e9182f8-81e1-4ca2-97b6-0066faba0b22" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.458s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.488261] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.572838] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc9cfc0c-4c8d-4f58-a1ee-a6363510a971 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.585022] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf69cab-ddaa-423f-a553-c6b8b5c9c43f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.616196] env[62235]: DEBUG nova.network.neutron [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 981.618673] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d4db33-690f-4ea0-86e9-2e59af37e46f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.627301] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd02571d-abab-482d-9918-32af0a438ff7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.642662] env[62235]: DEBUG nova.compute.provider_tree [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.704932] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.705267] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31265ef9-92b8-467f-8b01-632c0ede22e7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.714202] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 981.714202] env[62235]: value = "task-1272054" [ 981.714202] env[62235]: _type = "Task" [ 981.714202] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.724020] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.781805] env[62235]: DEBUG nova.network.neutron [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Updating instance_info_cache with network_info: [{"id": "511a3ca7-0c6b-4e8f-a157-143ce99d2382", "address": "fa:16:3e:27:11:aa", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap511a3ca7-0c", "ovs_interfaceid": "511a3ca7-0c6b-4e8f-a157-143ce99d2382", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.146304] env[62235]: DEBUG nova.scheduler.client.report [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.225088] env[62235]: DEBUG oslo_vmware.api [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272054, 'name': PowerOnVM_Task, 'duration_secs': 0.490412} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.225382] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 982.225570] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c6105890-2cd1-4e18-9971-d8f3b28eea3a tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance '3ea7db36-ebde-4523-b803-ad121a00269b' progress to 100 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.285384] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "refresh_cache-8a6962ba-4406-4ac9-8447-b839903781fe" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.285746] env[62235]: DEBUG nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Instance network_info: |[{"id": "511a3ca7-0c6b-4e8f-a157-143ce99d2382", "address": "fa:16:3e:27:11:aa", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap511a3ca7-0c", "ovs_interfaceid": "511a3ca7-0c6b-4e8f-a157-143ce99d2382", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.286265] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:11:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'df1bf911-aac9-4d2d-ae69-66ace3e6a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '511a3ca7-0c6b-4e8f-a157-143ce99d2382', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.294062] env[62235]: DEBUG oslo.service.loopingcall [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.294893] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.295163] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99b94ad4-330d-4959-8efb-b5dbf2cdfc65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.317257] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.317257] env[62235]: value = "task-1272055" [ 982.317257] env[62235]: _type = "Task" [ 982.317257] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.322546] env[62235]: INFO nova.compute.manager [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Rescuing [ 982.322822] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.322981] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.323169] env[62235]: DEBUG nova.network.neutron [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.327399] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272055, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.483190] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "interface-a3336b20-5737-4e5a-94a1-19fec140aa60-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.483573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-a3336b20-5737-4e5a-94a1-19fec140aa60-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.637487] env[62235]: DEBUG nova.compute.manager [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Received event network-changed-511a3ca7-0c6b-4e8f-a157-143ce99d2382 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.637727] env[62235]: DEBUG nova.compute.manager [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Refreshing instance network info cache due to event network-changed-511a3ca7-0c6b-4e8f-a157-143ce99d2382. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.637928] env[62235]: DEBUG oslo_concurrency.lockutils [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] Acquiring lock "refresh_cache-8a6962ba-4406-4ac9-8447-b839903781fe" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.638088] env[62235]: DEBUG oslo_concurrency.lockutils [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] Acquired lock "refresh_cache-8a6962ba-4406-4ac9-8447-b839903781fe" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.638266] env[62235]: DEBUG nova.network.neutron [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Refreshing network info cache for port 511a3ca7-0c6b-4e8f-a157-143ce99d2382 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.653207] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.725s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.655516] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.490s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.657670] env[62235]: INFO nova.compute.claims [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.686900] env[62235]: INFO nova.scheduler.client.report [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted allocations for instance 15d41c0f-045d-4e1c-88b5-1841f99f2e74 [ 982.834476] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272055, 'name': CreateVM_Task, 'duration_secs': 0.396531} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.834662] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.835568] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.835743] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.836074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.836339] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-353ea17c-2d5f-4b62-b06f-132783621b72 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.841232] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 982.841232] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520466b1-a155-73f4-e390-dbffd5221210" [ 982.841232] env[62235]: _type = "Task" [ 982.841232] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.849687] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520466b1-a155-73f4-e390-dbffd5221210, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.988957] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.988957] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.988957] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8adf154-49b3-41e4-9851-2170177aab79 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.016310] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37404665-8df3-4e7b-9c38-3d475b7cfbc2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.046514] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Reconfiguring VM to detach interface {{(pid=62235) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 983.046904] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2324670a-5aa1-4246-9203-dce23fadddba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.068920] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 983.068920] env[62235]: value = "task-1272056" [ 983.068920] env[62235]: _type = "Task" [ 983.068920] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.081264] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.137901] env[62235]: DEBUG nova.network.neutron [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Updating instance_info_cache with network_info: [{"id": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "address": "fa:16:3e:ec:f5:f9", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e813a1d-b6", "ovs_interfaceid": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.201643] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a64acd60-0fdd-4265-8073-bf542b9fb932 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "15d41c0f-045d-4e1c-88b5-1841f99f2e74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.201s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.354550] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520466b1-a155-73f4-e390-dbffd5221210, 'name': SearchDatastore_Task, 'duration_secs': 0.010731} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.354887] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.355143] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.355381] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.355532] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.355804] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.355983] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-705dfb40-1deb-4cdf-9dc1-313759de82c6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.366143] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.366337] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.367066] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95f39c70-1328-4c90-b4c0-562c3e1775b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.372849] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 983.372849] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2d3ff-fc78-7ccb-d83b-6636b3451cd9" [ 983.372849] env[62235]: _type = "Task" [ 983.372849] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.380703] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2d3ff-fc78-7ccb-d83b-6636b3451cd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.414738] env[62235]: DEBUG nova.network.neutron [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Updated VIF entry in instance network info cache for port 511a3ca7-0c6b-4e8f-a157-143ce99d2382. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.415132] env[62235]: DEBUG nova.network.neutron [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Updating instance_info_cache with network_info: [{"id": "511a3ca7-0c6b-4e8f-a157-143ce99d2382", "address": "fa:16:3e:27:11:aa", "network": {"id": "e3d22e5f-2c81-4c43-8a4a-bd22faa2081a", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-163866016-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4469550d32b4482a9d5e9244cbd681e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "df1bf911-aac9-4d2d-ae69-66ace3e6a2d1", "external-id": "nsx-vlan-transportzone-395", "segmentation_id": 395, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap511a3ca7-0c", "ovs_interfaceid": "511a3ca7-0c6b-4e8f-a157-143ce99d2382", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.579867] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.644585] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.805591] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9cb94a-88b9-4a58-a766-0def264f5b3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.814412] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f452d526-2c12-4e06-88f8-d9d86cef270c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.845562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0f45e3-8394-4030-bb1a-0afdcfa92499 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.853075] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c9d360-20cd-4525-86e2-1b82cc8c02d7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.866279] env[62235]: DEBUG nova.compute.provider_tree [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.883062] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b2d3ff-fc78-7ccb-d83b-6636b3451cd9, 'name': SearchDatastore_Task, 'duration_secs': 0.01198} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.883850] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b18726e-0352-4afa-bec2-8eb780211395 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.888814] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 983.888814] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f3b7e2-4bdc-f5be-6c8e-a21e1e232b6f" [ 983.888814] env[62235]: _type = "Task" [ 983.888814] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.897354] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f3b7e2-4bdc-f5be-6c8e-a21e1e232b6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.917347] env[62235]: DEBUG oslo_concurrency.lockutils [req-59218b57-aff8-451b-bc37-807c13e85b02 req-e672eb96-8be3-4271-bb5c-0c4016cc7634 service nova] Releasing lock "refresh_cache-8a6962ba-4406-4ac9-8447-b839903781fe" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.082579] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.104746] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.105376] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.181766] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.182078] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d1e4e23-eed4-4025-8c44-e059cc87de5a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.190468] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 984.190468] env[62235]: value = "task-1272057" [ 984.190468] env[62235]: _type = "Task" [ 984.190468] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.200127] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.368841] env[62235]: DEBUG nova.scheduler.client.report [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.403691] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f3b7e2-4bdc-f5be-6c8e-a21e1e232b6f, 'name': SearchDatastore_Task, 'duration_secs': 0.009619} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.404031] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.404477] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8a6962ba-4406-4ac9-8447-b839903781fe/8a6962ba-4406-4ac9-8447-b839903781fe.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.404817] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c2b2706-e0a2-41bd-b773-2cd277f4dd20 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.415534] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 984.415534] env[62235]: value = "task-1272058" [ 984.415534] env[62235]: _type = "Task" [ 984.415534] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.424198] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.533850] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.534267] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.534557] env[62235]: DEBUG nova.compute.manager [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Going to confirm migration 3 {{(pid=62235) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 984.584097] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.607773] env[62235]: DEBUG nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.700965] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272057, 'name': PowerOffVM_Task, 'duration_secs': 0.273473} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.701301] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.702177] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f99183-d4a9-4b7c-b673-e96e82589913 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.722663] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcac2b16-7932-4cb1-9dba-a8270c0c5c11 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.754618] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.755036] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3195aca-94eb-4576-850e-b762ad4d1629 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.763688] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 984.763688] env[62235]: value = "task-1272059" [ 984.763688] env[62235]: _type = "Task" [ 984.763688] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.773289] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.874077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.874611] env[62235]: DEBUG nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 984.877346] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.155s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.877586] env[62235]: DEBUG nova.objects.instance [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lazy-loading 'resources' on Instance uuid 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.926718] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272058, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464724} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.926992] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8a6962ba-4406-4ac9-8447-b839903781fe/8a6962ba-4406-4ac9-8447-b839903781fe.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 984.927224] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.927476] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e38c1fa6-b71f-42c3-a796-5e165034f27b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.934278] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 984.934278] env[62235]: value = "task-1272060" [ 984.934278] env[62235]: _type = "Task" [ 984.934278] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.941628] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272060, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.081689] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.126392] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.132191] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.132368] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.132540] env[62235]: DEBUG nova.network.neutron [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.132755] env[62235]: DEBUG nova.objects.instance [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'info_cache' on Instance uuid 3ea7db36-ebde-4523-b803-ad121a00269b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.241413] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "e98e81bd-0bde-4b23-b380-18544512dc20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.241649] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.274596] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 985.274838] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.275096] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.275255] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.275442] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.275687] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f874d0b4-eea3-4f92-9c5e-225f21132c67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.291984] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.292187] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.292957] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa47f6df-d826-4a0b-9c9a-59e99eb8e339 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.298249] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 985.298249] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52999738-4a0f-474a-9a3a-a2bbca14ea24" [ 985.298249] env[62235]: _type = "Task" [ 985.298249] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.305449] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52999738-4a0f-474a-9a3a-a2bbca14ea24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.380276] env[62235]: DEBUG nova.compute.utils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.385019] env[62235]: DEBUG nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.385019] env[62235]: DEBUG nova.network.neutron [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.420995] env[62235]: DEBUG nova.policy [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96af7a114d04c0dad3c53f5a2dcd2bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef6547723ccf4be891da50a179dae06f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.442396] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272060, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065227} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.442689] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.443469] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c2f830-60af-4bbf-a65b-5871ffcb2e8d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.466829] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 8a6962ba-4406-4ac9-8447-b839903781fe/8a6962ba-4406-4ac9-8447-b839903781fe.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.469285] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ae7b377-5db3-4b9e-8aec-dbf66e0204b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.490900] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 985.490900] env[62235]: value = "task-1272061" [ 985.490900] env[62235]: _type = "Task" [ 985.490900] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.500483] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272061, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.558716] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cabfaa-9c50-47c3-8099-f167a8b262e8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.566620] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d992e97-eff4-4f03-b1c0-0849953f48d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.599098] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4428e58-92ad-4206-bf62-8778365dcd56 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.606567] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.610122] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78984c27-e54b-4621-aba4-1f1d7b0c82de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.622496] env[62235]: DEBUG nova.compute.provider_tree [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.727974] env[62235]: DEBUG nova.network.neutron [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Successfully created port: 773479e6-82f4-4668-be25-fd24885c78ed {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.743889] env[62235]: DEBUG nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 985.808611] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52999738-4a0f-474a-9a3a-a2bbca14ea24, 'name': SearchDatastore_Task, 'duration_secs': 0.011027} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.809600] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cbebaea-7969-49fc-a255-35585b13dc25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.814715] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 985.814715] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526faaab-7fa9-f6c0-7a8c-804f22584687" [ 985.814715] env[62235]: _type = "Task" [ 985.814715] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.824561] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526faaab-7fa9-f6c0-7a8c-804f22584687, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.889103] env[62235]: DEBUG nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.000190] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272061, 'name': ReconfigVM_Task, 'duration_secs': 0.26257} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.000491] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 8a6962ba-4406-4ac9-8447-b839903781fe/8a6962ba-4406-4ac9-8447-b839903781fe.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.001129] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2a85549-fbdb-4dee-b781-5e974c775934 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.006870] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 986.006870] env[62235]: value = "task-1272062" [ 986.006870] env[62235]: _type = "Task" [ 986.006870] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.013836] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272062, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.104910] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.127104] env[62235]: DEBUG nova.scheduler.client.report [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.267893] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.326458] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526faaab-7fa9-f6c0-7a8c-804f22584687, 'name': SearchDatastore_Task, 'duration_secs': 0.008796} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.328904] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.329196] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk. {{(pid=62235) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 986.329469] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e5ba5dd-396d-469c-85f7-2b91721fa20e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.336666] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 986.336666] env[62235]: value = "task-1272063" [ 986.336666] env[62235]: _type = "Task" [ 986.336666] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.344142] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.365238] env[62235]: DEBUG nova.network.neutron [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [{"id": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "address": "fa:16:3e:31:cc:a3", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeed03567-8c", "ovs_interfaceid": "eed03567-8c20-4e56-902c-fcd6bb25b56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.517887] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272062, 'name': Rename_Task, 'duration_secs': 0.141219} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.517887] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 986.517887] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26fc5438-75e0-42f7-b295-5547a602d809 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.524733] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 986.524733] env[62235]: value = "task-1272064" [ 986.524733] env[62235]: _type = "Task" [ 986.524733] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.533107] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272064, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.605932] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.632115] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.635404] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.146s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.635633] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.635832] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 986.636340] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.510s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.638477] env[62235]: INFO nova.compute.claims [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.643568] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14195aa6-1707-49ba-a9db-5a74cfa04231 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.653906] env[62235]: INFO nova.scheduler.client.report [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleted allocations for instance 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf [ 986.656926] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb050c5-6a90-4f32-a9d7-004fd9a79a6a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.676727] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b100bfff-5fd9-4732-bcd3-d756553d30e4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.685754] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeeab0e6-9e9d-4f8a-8ab1-b5becc3b25fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.722132] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180567MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 986.722435] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.847263] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454456} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.847599] env[62235]: INFO nova.virt.vmwareapi.ds_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk. [ 986.848432] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817ab7bb-afcd-4f7b-af87-fc1fd21222b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.873507] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.874155] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-3ea7db36-ebde-4523-b803-ad121a00269b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.874392] env[62235]: DEBUG nova.objects.instance [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'migration_context' on Instance uuid 3ea7db36-ebde-4523-b803-ad121a00269b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.875481] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68e63a31-82b2-4cdc-aa9c-e1d33d126777 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.889817] env[62235]: DEBUG nova.objects.base [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Object Instance<3ea7db36-ebde-4523-b803-ad121a00269b> lazy-loaded attributes: info_cache,migration_context {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 986.890666] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cddb58-d86b-4ed1-8f56-31ed24404126 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.897410] env[62235]: DEBUG nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 986.914967] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 986.914967] env[62235]: value = "task-1272065" [ 986.914967] env[62235]: _type = "Task" [ 986.914967] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.915214] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-973483a1-feda-4215-b8e7-a55cdfcf18ca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.925064] env[62235]: DEBUG oslo_vmware.api [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 986.925064] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d04e77-5939-d0ce-4266-4171196897cb" [ 986.925064] env[62235]: _type = "Task" [ 986.925064] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.929034] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272065, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.942361] env[62235]: DEBUG oslo_vmware.api [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d04e77-5939-d0ce-4266-4171196897cb, 'name': SearchDatastore_Task, 'duration_secs': 0.007401} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.944581] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.944823] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.945030] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.945256] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.945422] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.945590] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.945860] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.946168] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.946414] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.946634] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.946880] env[62235]: DEBUG nova.virt.hardware [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.947296] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.948093] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837af3bf-e3ad-40ae-94d0-545ceda1e7ce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.955614] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6dff31-1b96-4292-9691-48fc25945d78 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.035838] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272064, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.105770] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.148523] env[62235]: DEBUG nova.compute.manager [req-0aef3d09-5102-43a3-abec-a6c9cd2589ec req-627c617a-1edc-40bc-9d16-7ed0831ec417 service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Received event network-vif-plugged-773479e6-82f4-4668-be25-fd24885c78ed {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.148878] env[62235]: DEBUG oslo_concurrency.lockutils [req-0aef3d09-5102-43a3-abec-a6c9cd2589ec req-627c617a-1edc-40bc-9d16-7ed0831ec417 service nova] Acquiring lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.149126] env[62235]: DEBUG oslo_concurrency.lockutils [req-0aef3d09-5102-43a3-abec-a6c9cd2589ec req-627c617a-1edc-40bc-9d16-7ed0831ec417 service nova] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.149648] env[62235]: DEBUG oslo_concurrency.lockutils [req-0aef3d09-5102-43a3-abec-a6c9cd2589ec req-627c617a-1edc-40bc-9d16-7ed0831ec417 service nova] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.149854] env[62235]: DEBUG nova.compute.manager [req-0aef3d09-5102-43a3-abec-a6c9cd2589ec req-627c617a-1edc-40bc-9d16-7ed0831ec417 service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] No waiting events found dispatching network-vif-plugged-773479e6-82f4-4668-be25-fd24885c78ed {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.149994] env[62235]: WARNING nova.compute.manager [req-0aef3d09-5102-43a3-abec-a6c9cd2589ec req-627c617a-1edc-40bc-9d16-7ed0831ec417 service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Received unexpected event network-vif-plugged-773479e6-82f4-4668-be25-fd24885c78ed for instance with vm_state building and task_state spawning. [ 987.168983] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d1dcf0d4-8b18-451f-b31f-dc0f1175fcf9 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.403s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.242960] env[62235]: DEBUG nova.network.neutron [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Successfully updated port: 773479e6-82f4-4668-be25-fd24885c78ed {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.428443] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272065, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.536760] env[62235]: DEBUG oslo_vmware.api [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272064, 'name': PowerOnVM_Task, 'duration_secs': 0.540813} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.537015] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 987.537237] env[62235]: INFO nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Took 7.74 seconds to spawn the instance on the hypervisor. [ 987.537424] env[62235]: DEBUG nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.538212] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49e0802-3136-42e5-92b2-4f250e290313 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.606157] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.745431] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-0071c426-0ed5-4e75-bf7f-cbf6221cda07" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.745595] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-0071c426-0ed5-4e75-bf7f-cbf6221cda07" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.745752] env[62235]: DEBUG nova.network.neutron [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.776452] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff46b029-f636-4093-b349-25cf198be781 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.783938] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110eaf35-0b98-4e13-afcd-19f2c683e0ff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.815136] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aaf0d53-fc50-44e2-a9c3-1f9c4578844f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.822034] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5621cc-28c3-49ec-ac67-e874a3fb5f0b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.834866] env[62235]: DEBUG nova.compute.provider_tree [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.928484] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272065, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.057045] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "eccde807-664e-4488-bec8-2d6919c2fc91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.057045] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.057045] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "eccde807-664e-4488-bec8-2d6919c2fc91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.057045] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.057045] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.059131] env[62235]: INFO nova.compute.manager [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Terminating instance [ 988.061553] env[62235]: INFO nova.compute.manager [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Took 18.69 seconds to build instance. [ 988.062770] env[62235]: DEBUG nova.compute.manager [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.065020] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.065020] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b562933-9457-4cb6-ba54-8d0a603f54e8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.072556] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.072953] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d384d9ff-3216-4250-815d-9faa03bb43cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.079069] env[62235]: DEBUG oslo_vmware.api [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 988.079069] env[62235]: value = "task-1272066" [ 988.079069] env[62235]: _type = "Task" [ 988.079069] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.087831] env[62235]: DEBUG oslo_vmware.api [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272066, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.106675] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.291642] env[62235]: DEBUG nova.network.neutron [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.337326] env[62235]: DEBUG nova.scheduler.client.report [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.430978] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272065, 'name': ReconfigVM_Task, 'duration_secs': 1.137362} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.431128] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f/02c64327-1eca-4d55-9f01-62fa8b2a2334-rescue.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.432015] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adcaaa8-f163-41a4-8a66-247a0b252871 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.459796] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22044c96-5089-4aa4-ba40-ae1813cdc8dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.474808] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 988.474808] env[62235]: value = "task-1272067" [ 988.474808] env[62235]: _type = "Task" [ 988.474808] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.482531] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272067, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.506463] env[62235]: DEBUG nova.network.neutron [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Updating instance_info_cache with network_info: [{"id": "773479e6-82f4-4668-be25-fd24885c78ed", "address": "fa:16:3e:2e:1d:13", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773479e6-82", "ovs_interfaceid": "773479e6-82f4-4668-be25-fd24885c78ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.568890] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2fc2455f-e1b6-4e5c-90ad-45d4f37efdb2 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "8a6962ba-4406-4ac9-8447-b839903781fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.206s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.589659] env[62235]: DEBUG oslo_vmware.api [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272066, 'name': PowerOffVM_Task, 'duration_secs': 0.170207} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.589946] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.590133] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.590395] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26427aac-46c2-459b-b782-dc8a8d3aa3eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.606925] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.669662] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 988.669894] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 988.670097] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleting the datastore file [datastore2] eccde807-664e-4488-bec8-2d6919c2fc91 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.670369] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82549d31-cd28-4f6d-9be2-323fcf833362 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.677394] env[62235]: DEBUG oslo_vmware.api [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for the task: (returnval){ [ 988.677394] env[62235]: value = "task-1272069" [ 988.677394] env[62235]: _type = "Task" [ 988.677394] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.685746] env[62235]: DEBUG oslo_vmware.api [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.843837] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.207s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.845029] env[62235]: DEBUG nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.847086] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.579s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.850030] env[62235]: INFO nova.compute.claims [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.985248] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272067, 'name': ReconfigVM_Task, 'duration_secs': 0.182407} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.985610] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.985810] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b26895c1-c037-4554-9221-3c140f91f427 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.992080] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 988.992080] env[62235]: value = "task-1272070" [ 988.992080] env[62235]: _type = "Task" [ 988.992080] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.999281] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272070, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.008823] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-0071c426-0ed5-4e75-bf7f-cbf6221cda07" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.009131] env[62235]: DEBUG nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Instance network_info: |[{"id": "773479e6-82f4-4668-be25-fd24885c78ed", "address": "fa:16:3e:2e:1d:13", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773479e6-82", "ovs_interfaceid": "773479e6-82f4-4668-be25-fd24885c78ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.009504] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:1d:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '773479e6-82f4-4668-be25-fd24885c78ed', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.017180] env[62235]: DEBUG oslo.service.loopingcall [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.017688] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.017913] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f67d7c67-add2-4ef6-87a3-5ee818e48ce2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.039447] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.039447] env[62235]: value = "task-1272071" [ 989.039447] env[62235]: _type = "Task" [ 989.039447] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.047173] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272071, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.112887] env[62235]: DEBUG oslo_vmware.api [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272056, 'name': ReconfigVM_Task, 'duration_secs': 5.773941} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.113208] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.113444] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Reconfigured VM to detach interface {{(pid=62235) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 989.178545] env[62235]: DEBUG nova.compute.manager [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Received event network-changed-773479e6-82f4-4668-be25-fd24885c78ed {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.178744] env[62235]: DEBUG nova.compute.manager [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Refreshing instance network info cache due to event network-changed-773479e6-82f4-4668-be25-fd24885c78ed. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.178976] env[62235]: DEBUG oslo_concurrency.lockutils [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] Acquiring lock "refresh_cache-0071c426-0ed5-4e75-bf7f-cbf6221cda07" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.179242] env[62235]: DEBUG oslo_concurrency.lockutils [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] Acquired lock "refresh_cache-0071c426-0ed5-4e75-bf7f-cbf6221cda07" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.179335] env[62235]: DEBUG nova.network.neutron [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Refreshing network info cache for port 773479e6-82f4-4668-be25-fd24885c78ed {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.192241] env[62235]: DEBUG oslo_vmware.api [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Task: {'id': task-1272069, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.411517} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.192518] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.192707] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 989.192914] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 989.193121] env[62235]: INFO nova.compute.manager [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Took 1.13 seconds to destroy the instance on the hypervisor. [ 989.193377] env[62235]: DEBUG oslo.service.loopingcall [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.193555] env[62235]: DEBUG nova.compute.manager [-] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.193646] env[62235]: DEBUG nova.network.neutron [-] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 989.316677] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "8a6962ba-4406-4ac9-8447-b839903781fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.316958] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "8a6962ba-4406-4ac9-8447-b839903781fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.317191] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "8a6962ba-4406-4ac9-8447-b839903781fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.317481] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "8a6962ba-4406-4ac9-8447-b839903781fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.317579] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "8a6962ba-4406-4ac9-8447-b839903781fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.319733] env[62235]: INFO nova.compute.manager [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Terminating instance [ 989.321619] env[62235]: DEBUG nova.compute.manager [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.321909] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.322653] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5335e2-2dc3-4ce2-a861-40d625c200ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.331043] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.331043] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fee29f0-3fd4-4013-808a-07f7131127c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.338097] env[62235]: DEBUG oslo_vmware.api [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 989.338097] env[62235]: value = "task-1272072" [ 989.338097] env[62235]: _type = "Task" [ 989.338097] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.346532] env[62235]: DEBUG oslo_vmware.api [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.353476] env[62235]: DEBUG nova.compute.utils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.357340] env[62235]: DEBUG nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.357602] env[62235]: DEBUG nova.network.neutron [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 989.465210] env[62235]: DEBUG nova.policy [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6018b2929a594c358fe543f4fa2d2f46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6832a93601f84138aeed5a906f24394b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.506304] env[62235]: DEBUG oslo_vmware.api [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272070, 'name': PowerOnVM_Task, 'duration_secs': 0.469035} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.506304] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.509756] env[62235]: DEBUG nova.compute.manager [None req-5827bc9a-6d61-4d84-92c0-2bd88733501f tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.510608] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6f4063-5d46-4157-ba34-049dde60d77e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.550810] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272071, 'name': CreateVM_Task, 'duration_secs': 0.359726} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.551031] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.551857] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.552108] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.552944] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.552944] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ef8c73d-bd2e-4b38-9d8a-3244ab501aaf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.557521] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 989.557521] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e7ffc3-53ff-0ac0-c0b3-4d430b5a61f7" [ 989.557521] env[62235]: _type = "Task" [ 989.557521] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.568664] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e7ffc3-53ff-0ac0-c0b3-4d430b5a61f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.617248] env[62235]: DEBUG nova.compute.manager [req-2126eb76-da52-49a4-9423-21b0050b17dc req-016f5669-9414-47dd-9c50-382ed1e8588b service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Received event network-vif-deleted-f505e074-0353-4a28-abad-d058112f253a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.617518] env[62235]: INFO nova.compute.manager [req-2126eb76-da52-49a4-9423-21b0050b17dc req-016f5669-9414-47dd-9c50-382ed1e8588b service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Neutron deleted interface f505e074-0353-4a28-abad-d058112f253a; detaching it from the instance and deleting it from the info cache [ 989.617720] env[62235]: DEBUG nova.network.neutron [req-2126eb76-da52-49a4-9423-21b0050b17dc req-016f5669-9414-47dd-9c50-382ed1e8588b service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.846813] env[62235]: DEBUG oslo_vmware.api [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272072, 'name': PowerOffVM_Task, 'duration_secs': 0.198309} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.847099] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.847274] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.847531] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75f42e2a-d5fe-4653-b9cd-31d5afe85f54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.859159] env[62235]: DEBUG nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.917547] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.917853] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.918066] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleting the datastore file [datastore1] 8a6962ba-4406-4ac9-8447-b839903781fe {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.918622] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d7886b9-8555-49cd-ab21-2010828805d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.929449] env[62235]: DEBUG oslo_vmware.api [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for the task: (returnval){ [ 989.929449] env[62235]: value = "task-1272074" [ 989.929449] env[62235]: _type = "Task" [ 989.929449] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.938079] env[62235]: DEBUG oslo_vmware.api [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.023190] env[62235]: DEBUG nova.network.neutron [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Successfully created port: 808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.069345] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e7ffc3-53ff-0ac0-c0b3-4d430b5a61f7, 'name': SearchDatastore_Task, 'duration_secs': 0.010081} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.069713] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.069946] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.070216] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.070371] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.070572] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.071098] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7949088a-c8fc-4fcb-8f5d-ee3e8ecba557 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.074293] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1eca93a-dfee-4a0e-9506-27e3a8aa7052 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.082610] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b71abbd-3e3c-44f0-a9bc-87d149b225ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.086598] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.086782] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.087815] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-804eb0c6-648f-4632-b72f-4b911d1a2c44 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.116979] env[62235]: DEBUG nova.network.neutron [-] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.121555] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936734b5-26fa-4600-8abb-b45f1c9cc710 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.125016] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 990.125016] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521518c2-85ba-1de3-1c46-b3415ade2be0" [ 990.125016] env[62235]: _type = "Task" [ 990.125016] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.125016] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ebeb7bfc-f1f7-410f-9db0-15bc9a1e3955 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.133134] env[62235]: DEBUG nova.network.neutron [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Updated VIF entry in instance network info cache for port 773479e6-82f4-4668-be25-fd24885c78ed. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.133505] env[62235]: DEBUG nova.network.neutron [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Updating instance_info_cache with network_info: [{"id": "773479e6-82f4-4668-be25-fd24885c78ed", "address": "fa:16:3e:2e:1d:13", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773479e6-82", "ovs_interfaceid": "773479e6-82f4-4668-be25-fd24885c78ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.136689] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a32a94-ea55-49a8-a6d6-8c7a7ab8aeb1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.148043] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521518c2-85ba-1de3-1c46-b3415ade2be0, 'name': SearchDatastore_Task, 'duration_secs': 0.013439} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.151550] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b688405e-ea0f-4cf9-a514-4a22299488d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.169777] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f58c49b-9e22-40d8-919f-925201576162 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.172291] env[62235]: DEBUG nova.compute.provider_tree [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.177660] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 990.177660] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520e8433-a67a-1104-6ebe-8e9a0dc386af" [ 990.177660] env[62235]: _type = "Task" [ 990.177660] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.186304] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520e8433-a67a-1104-6ebe-8e9a0dc386af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.193637] env[62235]: DEBUG nova.compute.manager [req-2126eb76-da52-49a4-9423-21b0050b17dc req-016f5669-9414-47dd-9c50-382ed1e8588b service nova] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Detach interface failed, port_id=f505e074-0353-4a28-abad-d058112f253a, reason: Instance eccde807-664e-4488-bec8-2d6919c2fc91 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 990.439247] env[62235]: DEBUG oslo_vmware.api [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Task: {'id': task-1272074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193555} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.439527] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.439713] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.439899] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.440088] env[62235]: INFO nova.compute.manager [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Took 1.12 seconds to destroy the instance on the hypervisor. [ 990.440330] env[62235]: DEBUG oslo.service.loopingcall [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.440521] env[62235]: DEBUG nova.compute.manager [-] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.440617] env[62235]: DEBUG nova.network.neutron [-] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.455575] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.455746] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquired lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.455920] env[62235]: DEBUG nova.network.neutron [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.457940] env[62235]: INFO nova.compute.manager [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Unrescuing [ 990.458313] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.458485] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquired lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.458660] env[62235]: DEBUG nova.network.neutron [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.619638] env[62235]: INFO nova.compute.manager [-] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Took 1.43 seconds to deallocate network for instance. [ 990.636424] env[62235]: DEBUG oslo_concurrency.lockutils [req-27400972-aca7-4d3b-aeb8-7a297922a223 req-c708cd45-4766-4b28-b596-5ceebf2d978e service nova] Releasing lock "refresh_cache-0071c426-0ed5-4e75-bf7f-cbf6221cda07" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.675308] env[62235]: DEBUG nova.scheduler.client.report [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.688761] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520e8433-a67a-1104-6ebe-8e9a0dc386af, 'name': SearchDatastore_Task, 'duration_secs': 0.009561} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.689580] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.690276] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 0071c426-0ed5-4e75-bf7f-cbf6221cda07/0071c426-0ed5-4e75-bf7f-cbf6221cda07.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.690276] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1150d8b6-7d45-4b38-a2d6-eec61cdd48eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.698035] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 990.698035] env[62235]: value = "task-1272075" [ 990.698035] env[62235]: _type = "Task" [ 990.698035] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.707858] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.871540] env[62235]: DEBUG nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.900148] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.900440] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.900604] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.900819] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.900986] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.901160] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.901371] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.901545] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.901751] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.901924] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.902111] env[62235]: DEBUG nova.virt.hardware [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.903084] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-005b0e9f-345f-45eb-a9f0-b208be99ffa7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.911618] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170a6bf3-ca8d-48d8-b551-c76178cf1e7b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.098411] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "a3336b20-5737-4e5a-94a1-19fec140aa60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.099257] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.099495] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.099700] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.100167] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.102868] env[62235]: INFO nova.compute.manager [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Terminating instance [ 991.105202] env[62235]: DEBUG nova.compute.manager [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.105451] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.106425] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9f03b2-6e17-4784-9288-972118b0643d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.117160] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.117160] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8caccd66-4b9e-4cd7-980d-3ace4d11404a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.122395] env[62235]: DEBUG oslo_vmware.api [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 991.122395] env[62235]: value = "task-1272076" [ 991.122395] env[62235]: _type = "Task" [ 991.122395] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.126314] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.134065] env[62235]: DEBUG oslo_vmware.api [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272076, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.160559] env[62235]: DEBUG nova.network.neutron [-] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.183290] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.336s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.184084] env[62235]: DEBUG nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.186998] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.465s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.214091] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412584} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.214611] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 0071c426-0ed5-4e75-bf7f-cbf6221cda07/0071c426-0ed5-4e75-bf7f-cbf6221cda07.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.215385] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.217901] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ba21f4b-5263-472f-a511-05323d074372 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.227979] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 991.227979] env[62235]: value = "task-1272077" [ 991.227979] env[62235]: _type = "Task" [ 991.227979] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.237396] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.277302] env[62235]: INFO nova.network.neutron [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Port ac6a259e-ebe5-4b85-b07f-38fe6b16c73b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 991.277733] env[62235]: DEBUG nova.network.neutron [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [{"id": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "address": "fa:16:3e:95:0d:82", "network": {"id": "695860c9-9102-4053-ad1c-75a1c4ac4b12", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-329313648-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1e885dc2e5404662a081bb3855919b0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bc70ab7-ba", "ovs_interfaceid": "1bc70ab7-baf7-4b1d-b945-41463b6ea517", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.485435] env[62235]: DEBUG nova.network.neutron [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Updating instance_info_cache with network_info: [{"id": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "address": "fa:16:3e:ec:f5:f9", "network": {"id": "ee92bfab-2d6c-4840-97cc-09dd5a6e473a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1717523802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b690e5ef3dac4a07879b4b7fe2249e89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e813a1d-b6", "ovs_interfaceid": "3e813a1d-b62c-400a-a376-ec445a16c4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.633924] env[62235]: DEBUG oslo_vmware.api [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272076, 'name': PowerOffVM_Task, 'duration_secs': 0.204197} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.634265] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.634448] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 991.634712] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5b596f9-7c4f-4dd1-8894-0f3a34290794 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.645817] env[62235]: DEBUG nova.compute.manager [req-004cf198-95c5-41a9-9ef0-5f0edf1f4b94 req-268b8c6c-d839-4af0-b826-de15d4b2ecfe service nova] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Received event network-vif-deleted-511a3ca7-0c6b-4e8f-a157-143ce99d2382 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.646034] env[62235]: DEBUG nova.compute.manager [req-004cf198-95c5-41a9-9ef0-5f0edf1f4b94 req-268b8c6c-d839-4af0-b826-de15d4b2ecfe service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-vif-plugged-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.646240] env[62235]: DEBUG oslo_concurrency.lockutils [req-004cf198-95c5-41a9-9ef0-5f0edf1f4b94 req-268b8c6c-d839-4af0-b826-de15d4b2ecfe service nova] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.646457] env[62235]: DEBUG oslo_concurrency.lockutils [req-004cf198-95c5-41a9-9ef0-5f0edf1f4b94 req-268b8c6c-d839-4af0-b826-de15d4b2ecfe service nova] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.646623] env[62235]: DEBUG oslo_concurrency.lockutils [req-004cf198-95c5-41a9-9ef0-5f0edf1f4b94 req-268b8c6c-d839-4af0-b826-de15d4b2ecfe service nova] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.646786] env[62235]: DEBUG nova.compute.manager [req-004cf198-95c5-41a9-9ef0-5f0edf1f4b94 req-268b8c6c-d839-4af0-b826-de15d4b2ecfe service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] No waiting events found dispatching network-vif-plugged-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.646952] env[62235]: WARNING nova.compute.manager [req-004cf198-95c5-41a9-9ef0-5f0edf1f4b94 req-268b8c6c-d839-4af0-b826-de15d4b2ecfe service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received unexpected event network-vif-plugged-808e6de1-f745-433c-b5e8-705d1fb7466b for instance with vm_state building and task_state spawning. [ 991.663459] env[62235]: DEBUG nova.network.neutron [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Successfully updated port: 808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.666128] env[62235]: INFO nova.compute.manager [-] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Took 1.23 seconds to deallocate network for instance. [ 991.691856] env[62235]: DEBUG nova.compute.utils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.699839] env[62235]: DEBUG nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.700064] env[62235]: DEBUG nova.network.neutron [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.702756] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 991.703008] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 991.703270] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleting the datastore file [datastore2] a3336b20-5737-4e5a-94a1-19fec140aa60 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.703727] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d94ac1f-8584-481f-a7a1-9db7c616b8c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.713028] env[62235]: DEBUG oslo_vmware.api [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 991.713028] env[62235]: value = "task-1272079" [ 991.713028] env[62235]: _type = "Task" [ 991.713028] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.722533] env[62235]: DEBUG oslo_vmware.api [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272079, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.736266] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071977} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.736565] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.737422] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f350792-e671-4e1d-80a4-61bd98c2bdf4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.760068] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 0071c426-0ed5-4e75-bf7f-cbf6221cda07/0071c426-0ed5-4e75-bf7f-cbf6221cda07.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.761623] env[62235]: DEBUG nova.policy [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8a1c9e076304498999fe4c6c1d89c69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a272872fcc55419fb474121b7fefd4c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.763527] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c28e3dc-4171-4b54-bbc9-5d252673c14c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.781963] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Releasing lock "refresh_cache-a3336b20-5737-4e5a-94a1-19fec140aa60" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.785302] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 991.785302] env[62235]: value = "task-1272080" [ 991.785302] env[62235]: _type = "Task" [ 991.785302] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.793639] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272080, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.988816] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Releasing lock "refresh_cache-550945fe-52be-408d-9f42-f5cd5e9c489f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.989568] env[62235]: DEBUG nova.objects.instance [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lazy-loading 'flavor' on Instance uuid 550945fe-52be-408d-9f42-f5cd5e9c489f {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.096201] env[62235]: DEBUG nova.network.neutron [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Successfully created port: 1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.166226] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.166392] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.166530] env[62235]: DEBUG nova.network.neutron [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 992.173063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.201250] env[62235]: DEBUG nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.204354] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Applying migration context for instance 3ea7db36-ebde-4523-b803-ad121a00269b as it has an incoming, in-progress migration 58fa497c-b13d-4311-b06e-5c36066eb749. Migration status is confirming {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 992.205534] env[62235]: INFO nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating resource usage from migration 58fa497c-b13d-4311-b06e-5c36066eb749 [ 992.227840] env[62235]: DEBUG oslo_vmware.api [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272079, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148085} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.228825] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.228825] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.228825] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.228825] env[62235]: INFO nova.compute.manager [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Took 1.12 seconds to destroy the instance on the hypervisor. [ 992.229140] env[62235]: DEBUG oslo.service.loopingcall [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.231166] env[62235]: DEBUG nova.compute.manager [-] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.231166] env[62235]: DEBUG nova.network.neutron [-] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.233467] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 2068403d-5870-4daa-b4d6-acaf39209488 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.233606] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a3336b20-5737-4e5a-94a1-19fec140aa60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.233751] env[62235]: WARNING nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance eccde807-664e-4488-bec8-2d6919c2fc91 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 992.233879] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance c47ea708-9a8b-44e6-a630-c20cfcd6f004 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.233995] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Migration 58fa497c-b13d-4311-b06e-5c36066eb749 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 992.234125] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 3ea7db36-ebde-4523-b803-ad121a00269b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.234241] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 550945fe-52be-408d-9f42-f5cd5e9c489f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.234353] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 8a6962ba-4406-4ac9-8447-b839903781fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.234463] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0071c426-0ed5-4e75-bf7f-cbf6221cda07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.234573] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 738750ad-931e-45e4-b0c6-8232fe1b9f46 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.234681] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance e98e81bd-0bde-4b23-b380-18544512dc20 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.234870] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 992.235013] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 992.288102] env[62235]: DEBUG oslo_concurrency.lockutils [None req-a3699bed-2268-4307-bcc8-145f739e504a tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "interface-a3336b20-5737-4e5a-94a1-19fec140aa60-ac6a259e-ebe5-4b85-b07f-38fe6b16c73b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.803s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.303321] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272080, 'name': ReconfigVM_Task, 'duration_secs': 0.266005} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.306839] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 0071c426-0ed5-4e75-bf7f-cbf6221cda07/0071c426-0ed5-4e75-bf7f-cbf6221cda07.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.310887] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-364ead5e-e4bd-49d7-b336-79e252471940 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.316654] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 992.316654] env[62235]: value = "task-1272081" [ 992.316654] env[62235]: _type = "Task" [ 992.316654] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.332128] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272081, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.428537] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5509a9-3447-489f-88d3-2d00c3a02579 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.436734] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec17fe1c-a87a-4207-abd9-02991c6fabc7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.469684] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e1fcb7-fdd1-4f40-912b-5f79dfaa173b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.477513] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6f01d6-bcd4-4e3f-8032-b01b7ae35f9d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.491463] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 992.497383] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d20919f-a2f9-4b3b-b2b7-009357b4183c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.520235] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.520568] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a6bf010-bd0b-4c2c-88b9-20a047b00468 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.528497] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 992.528497] env[62235]: value = "task-1272082" [ 992.528497] env[62235]: _type = "Task" [ 992.528497] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.536957] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.698725] env[62235]: DEBUG nova.network.neutron [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 992.826683] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272081, 'name': Rename_Task, 'duration_secs': 0.446133} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.826962] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.827219] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-617e75ff-def6-4081-a88f-6b47804679bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.834513] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 992.834513] env[62235]: value = "task-1272083" [ 992.834513] env[62235]: _type = "Task" [ 992.834513] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.841906] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.842826] env[62235]: DEBUG nova.network.neutron [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap808e6de1-f7", "ovs_interfaceid": "808e6de1-f745-433c-b5e8-705d1fb7466b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.015202] env[62235]: ERROR nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [req-5bfa087c-e123-4ac8-a529-19db40212f92] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 108a4390-ff68-4048-b61d-c7a9614ddc4f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5bfa087c-e123-4ac8-a529-19db40212f92"}]} [ 993.033235] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 993.042859] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272082, 'name': PowerOffVM_Task, 'duration_secs': 0.298644} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.043251] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.050593] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 993.051210] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 993.051557] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 993.053629] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f5853da-c898-4c37-b775-8fc2397da973 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.072776] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 993.072776] env[62235]: value = "task-1272084" [ 993.072776] env[62235]: _type = "Task" [ 993.072776] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.083435] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272084, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.084866] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 993.107763] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 993.142680] env[62235]: DEBUG nova.compute.manager [req-42d24fe8-f2b0-48ee-8464-b1b19ca3caed req-9b1d8915-03a8-44e6-9a70-d17cfcb12e48 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Received event network-vif-deleted-1bc70ab7-baf7-4b1d-b945-41463b6ea517 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.143081] env[62235]: INFO nova.compute.manager [req-42d24fe8-f2b0-48ee-8464-b1b19ca3caed req-9b1d8915-03a8-44e6-9a70-d17cfcb12e48 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Neutron deleted interface 1bc70ab7-baf7-4b1d-b945-41463b6ea517; detaching it from the instance and deleting it from the info cache [ 993.143163] env[62235]: DEBUG nova.network.neutron [req-42d24fe8-f2b0-48ee-8464-b1b19ca3caed req-9b1d8915-03a8-44e6-9a70-d17cfcb12e48 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.218222] env[62235]: DEBUG nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.251450] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.251738] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.251940] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.252180] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.252371] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.252559] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.252800] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.253019] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.253248] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.253451] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.253662] env[62235]: DEBUG nova.virt.hardware [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.254817] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994f7471-cf55-4b53-9044-7fa7ca503e58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.263379] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04d3410-cc15-468a-b5a1-a86a355cabe6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.284963] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be93895-dc93-49ad-9ccb-2cc6be35da29 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.292954] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52d2126-cb36-4ab7-a2c9-a60cb91d37e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.325417] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cda89f-df04-4608-9e7d-c16d94265aee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.333751] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3599829-60e8-490a-b895-863c77196f3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.345356] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.345869] env[62235]: DEBUG nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Instance network_info: |[{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap808e6de1-f7", "ovs_interfaceid": "808e6de1-f745-433c-b5e8-705d1fb7466b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 993.358060] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:c5:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '808e6de1-f745-433c-b5e8-705d1fb7466b', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.365403] env[62235]: DEBUG oslo.service.loopingcall [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.365793] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 993.367060] env[62235]: DEBUG oslo_vmware.api [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272083, 'name': PowerOnVM_Task, 'duration_secs': 0.470713} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.367520] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 993.367778] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.368007] env[62235]: INFO nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Took 6.47 seconds to spawn the instance on the hypervisor. [ 993.368217] env[62235]: DEBUG nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.368460] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d549d9c-0fac-49c6-8fae-7824c9d58970 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.386247] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdee0a9-c38c-40a4-a128-fec50a936d56 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.396724] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.396724] env[62235]: value = "task-1272085" [ 993.396724] env[62235]: _type = "Task" [ 993.396724] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.404389] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272085, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.583323] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272084, 'name': ReconfigVM_Task, 'duration_secs': 0.22923} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.583614] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 993.583801] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.584088] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd021227-afa2-4c82-b97d-fd8a1dfb61d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.591040] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 993.591040] env[62235]: value = "task-1272086" [ 993.591040] env[62235]: _type = "Task" [ 993.591040] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.598371] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.607029] env[62235]: DEBUG nova.network.neutron [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Successfully updated port: 1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.623240] env[62235]: DEBUG nova.network.neutron [-] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.647272] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07658e9b-c403-4945-ba81-49cca49ce515 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.657157] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6caca9c4-688e-4e2d-9f29-a3c02e70d1b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.682860] env[62235]: DEBUG nova.compute.manager [req-42d24fe8-f2b0-48ee-8464-b1b19ca3caed req-9b1d8915-03a8-44e6-9a70-d17cfcb12e48 service nova] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Detach interface failed, port_id=1bc70ab7-baf7-4b1d-b945-41463b6ea517, reason: Instance a3336b20-5737-4e5a-94a1-19fec140aa60 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 993.766058] env[62235]: DEBUG nova.compute.manager [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.766058] env[62235]: DEBUG nova.compute.manager [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing instance network info cache due to event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.766058] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.766058] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.767168] env[62235]: DEBUG nova.network.neutron [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.894786] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Updated inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 993.895027] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Updating resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f generation from 117 to 118 during operation: update_inventory {{(pid=62235) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 993.895181] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 993.909315] env[62235]: INFO nova.compute.manager [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Took 15.76 seconds to build instance. [ 993.915824] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272085, 'name': CreateVM_Task, 'duration_secs': 0.356594} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.916646] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.917310] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.917479] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.917854] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.918315] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d70902a0-446c-481e-87f3-5faf2a177fb0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.923203] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 993.923203] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521e2e4f-b4de-7f7b-d019-fd0dc25200a4" [ 993.923203] env[62235]: _type = "Task" [ 993.923203] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.596509] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.596680] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.596786] env[62235]: DEBUG nova.network.neutron [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 994.597839] env[62235]: INFO nova.compute.manager [-] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Took 2.37 seconds to deallocate network for instance. [ 994.600142] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 994.600320] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.413s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.600653] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0964949d-f1ad-425b-8853-0d41cf14b2ea tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.459s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.607759] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.660s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.617070] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521e2e4f-b4de-7f7b-d019-fd0dc25200a4, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.617609] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.617842] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.618147] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.618352] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.618590] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.621597] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c608d31-07ff-457d-b871-d46289305bc0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.624600] env[62235]: DEBUG oslo_vmware.api [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272086, 'name': PowerOnVM_Task, 'duration_secs': 0.952579} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.624848] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.625093] env[62235]: DEBUG nova.compute.manager [None req-e440a3b0-911b-465c-a03d-ee30440252dd tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.626198] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d96f3b-6491-43b6-9a00-7519f650c98d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.633143] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.633264] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.635944] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04eae655-5d80-4457-b021-bc55d7c4dd8f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.641448] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 994.641448] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f05bf3-225b-6e95-4c91-b9b1baad2112" [ 994.641448] env[62235]: _type = "Task" [ 994.641448] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.650056] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f05bf3-225b-6e95-4c91-b9b1baad2112, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.047642] env[62235]: DEBUG nova.network.neutron [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updated VIF entry in instance network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.047959] env[62235]: DEBUG nova.network.neutron [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap808e6de1-f7", "ovs_interfaceid": "808e6de1-f745-433c-b5e8-705d1fb7466b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.115385] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f164fa-ce07-4106-a4a8-a7cc1287a851 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.118690] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.123651] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9c6d5fb-db9a-42e7-a737-c5c5005c11c0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Suspending the VM {{(pid=62235) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 995.123936] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-8d191e60-6967-4ed6-89c3-e9661c085de4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.132820] env[62235]: DEBUG oslo_vmware.api [None req-f9c6d5fb-db9a-42e7-a737-c5c5005c11c0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 995.132820] env[62235]: value = "task-1272087" [ 995.132820] env[62235]: _type = "Task" [ 995.132820] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.133942] env[62235]: DEBUG nova.network.neutron [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 995.151737] env[62235]: DEBUG oslo_vmware.api [None req-f9c6d5fb-db9a-42e7-a737-c5c5005c11c0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272087, 'name': SuspendVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.157613] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f05bf3-225b-6e95-4c91-b9b1baad2112, 'name': SearchDatastore_Task, 'duration_secs': 0.008084} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.163071] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db4d8060-2891-4cb9-b63a-07ce4f4bbe80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.169099] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 995.169099] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ffcce8-d70d-2e57-5445-56bdeec1e890" [ 995.169099] env[62235]: _type = "Task" [ 995.169099] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.178083] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ffcce8-d70d-2e57-5445-56bdeec1e890, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.292165] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51195072-c0ff-41b3-861b-cfe2c6d31421 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.302334] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46db5073-9848-4fa8-9697-d616c1fa7924 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.334971] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723662c1-7f33-480a-827c-3b0689625ffc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.342942] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e864cc16-8e90-4a57-a859-df26fd38c04b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.356593] env[62235]: DEBUG nova.compute.provider_tree [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.377601] env[62235]: DEBUG nova.network.neutron [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updating instance_info_cache with network_info: [{"id": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "address": "fa:16:3e:aa:39:fb", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b0d00c1-f3", "ovs_interfaceid": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.551222] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.551514] env[62235]: DEBUG nova.compute.manager [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Received event network-vif-plugged-1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.551724] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Acquiring lock "e98e81bd-0bde-4b23-b380-18544512dc20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.551934] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Lock "e98e81bd-0bde-4b23-b380-18544512dc20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.552115] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Lock "e98e81bd-0bde-4b23-b380-18544512dc20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.552349] env[62235]: DEBUG nova.compute.manager [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] No waiting events found dispatching network-vif-plugged-1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.552461] env[62235]: WARNING nova.compute.manager [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Received unexpected event network-vif-plugged-1b0d00c1-f3f2-49d0-b742-9c083a6176fa for instance with vm_state building and task_state spawning. [ 995.552628] env[62235]: DEBUG nova.compute.manager [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Received event network-changed-1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.552786] env[62235]: DEBUG nova.compute.manager [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Refreshing instance network info cache due to event network-changed-1b0d00c1-f3f2-49d0-b742-9c083a6176fa. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.552957] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Acquiring lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.646559] env[62235]: DEBUG oslo_vmware.api [None req-f9c6d5fb-db9a-42e7-a737-c5c5005c11c0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272087, 'name': SuspendVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.679452] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ffcce8-d70d-2e57-5445-56bdeec1e890, 'name': SearchDatastore_Task, 'duration_secs': 0.017099} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.679711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.679950] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.680221] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56ced668-de86-4a38-bc1d-3869fbeffa31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.686931] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 995.686931] env[62235]: value = "task-1272088" [ 995.686931] env[62235]: _type = "Task" [ 995.686931] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.694085] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.859907] env[62235]: DEBUG nova.scheduler.client.report [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.880209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.880525] env[62235]: DEBUG nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Instance network_info: |[{"id": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "address": "fa:16:3e:aa:39:fb", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b0d00c1-f3", "ovs_interfaceid": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.880912] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Acquired lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.881186] env[62235]: DEBUG nova.network.neutron [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Refreshing network info cache for port 1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.882457] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:39:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b0d00c1-f3f2-49d0-b742-9c083a6176fa', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.891034] env[62235]: DEBUG oslo.service.loopingcall [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.891229] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 995.891511] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91651f60-b98a-4647-a1d8-6cc44cb42190 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.913375] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.913375] env[62235]: value = "task-1272089" [ 995.913375] env[62235]: _type = "Task" [ 995.913375] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.923925] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272089, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.146266] env[62235]: DEBUG oslo_vmware.api [None req-f9c6d5fb-db9a-42e7-a737-c5c5005c11c0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272087, 'name': SuspendVM_Task, 'duration_secs': 0.653602} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.146585] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f9c6d5fb-db9a-42e7-a737-c5c5005c11c0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Suspended the VM {{(pid=62235) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 996.146776] env[62235]: DEBUG nova.compute.manager [None req-f9c6d5fb-db9a-42e7-a737-c5c5005c11c0 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.147528] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfffae3-2ea4-4e90-a9e1-4a7818ea7989 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.198058] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440943} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.198374] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.198594] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.198903] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92b93288-29d5-498f-a00f-8bbab0be9f32 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.205748] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 996.205748] env[62235]: value = "task-1272090" [ 996.205748] env[62235]: _type = "Task" [ 996.205748] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.213611] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272090, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.422866] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272089, 'name': CreateVM_Task, 'duration_secs': 0.398189} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.423125] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.423897] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.424120] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.424466] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.424719] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d1627b9-d730-42b4-bc5b-2990fd096fcf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.429386] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 996.429386] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c9dfa0-ad2a-b97f-ca30-fa4d082dec87" [ 996.429386] env[62235]: _type = "Task" [ 996.429386] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.439352] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c9dfa0-ad2a-b97f-ca30-fa4d082dec87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.662161] env[62235]: DEBUG nova.network.neutron [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updated VIF entry in instance network info cache for port 1b0d00c1-f3f2-49d0-b742-9c083a6176fa. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 996.662516] env[62235]: DEBUG nova.network.neutron [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updating instance_info_cache with network_info: [{"id": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "address": "fa:16:3e:aa:39:fb", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b0d00c1-f3", "ovs_interfaceid": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.715154] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272090, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088976} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.715511] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.716250] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e089591e-e225-4d9b-b8ec-556f2baa9269 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.718779] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "550945fe-52be-408d-9f42-f5cd5e9c489f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.719073] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.719228] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "550945fe-52be-408d-9f42-f5cd5e9c489f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.719418] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.719589] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.721936] env[62235]: INFO nova.compute.manager [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Terminating instance [ 996.741702] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.742421] env[62235]: DEBUG nova.compute.manager [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.742626] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.743172] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a723fc97-0195-4e1a-8fb0-8790c96b137a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.757543] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0210062-9424-4e1d-a9df-06ba074af0a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.765390] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.766519] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-488022fa-7796-4785-b092-d87563311f5b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.767913] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 996.767913] env[62235]: value = "task-1272091" [ 996.767913] env[62235]: _type = "Task" [ 996.767913] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.772206] env[62235]: DEBUG oslo_vmware.api [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 996.772206] env[62235]: value = "task-1272092" [ 996.772206] env[62235]: _type = "Task" [ 996.772206] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.778229] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272091, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.783155] env[62235]: DEBUG oslo_vmware.api [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.870533] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.263s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.873583] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.747s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.873778] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.875796] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.703s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.876015] env[62235]: DEBUG nova.objects.instance [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lazy-loading 'resources' on Instance uuid 8a6962ba-4406-4ac9-8447-b839903781fe {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.892812] env[62235]: INFO nova.scheduler.client.report [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Deleted allocations for instance eccde807-664e-4488-bec8-2d6919c2fc91 [ 996.940941] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c9dfa0-ad2a-b97f-ca30-fa4d082dec87, 'name': SearchDatastore_Task, 'duration_secs': 0.00964} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.941452] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.941732] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.942010] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.942180] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.942387] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.942731] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69434ae6-2f51-43d7-8ed4-bccca1c1ec06 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.952984] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.953528] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.953983] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5cfdcf7-3283-44ac-8603-6f1ec8bb28b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.959918] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 996.959918] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525474e0-a874-7786-0475-061edf3f99c6" [ 996.959918] env[62235]: _type = "Task" [ 996.959918] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.969114] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525474e0-a874-7786-0475-061edf3f99c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.997810] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.998539] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.998925] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.999362] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.999574] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.001822] env[62235]: INFO nova.compute.manager [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Terminating instance [ 997.004133] env[62235]: DEBUG nova.compute.manager [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 997.004371] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.005223] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ea7067-9e61-447d-92f3-56b5e65755a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.012240] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.012480] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b76b59cf-c6dd-4ef6-beb3-aed28dfa50af {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.165040] env[62235]: DEBUG oslo_concurrency.lockutils [req-bbae2503-1e60-4ebb-83f4-ec35c8aef9fa req-2a989ded-5624-429d-b436-a01587f60b83 service nova] Releasing lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.278938] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272091, 'name': ReconfigVM_Task, 'duration_secs': 0.271717} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.279566] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.280211] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a172796b-fd9e-4cea-8d50-1c7956a59e24 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.284309] env[62235]: DEBUG oslo_vmware.api [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272092, 'name': PowerOffVM_Task, 'duration_secs': 0.176092} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.284849] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.285063] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.285291] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b274ea8-eabe-4694-9c9e-be27e36da48e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.289182] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 997.289182] env[62235]: value = "task-1272094" [ 997.289182] env[62235]: _type = "Task" [ 997.289182] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.296729] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272094, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.354993] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.355274] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.355484] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Deleting the datastore file [datastore2] 550945fe-52be-408d-9f42-f5cd5e9c489f {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.355783] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad304e12-04a4-4e91-93b8-d274067a49d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.362040] env[62235]: DEBUG oslo_vmware.api [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 997.362040] env[62235]: value = "task-1272096" [ 997.362040] env[62235]: _type = "Task" [ 997.362040] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.370018] env[62235]: DEBUG oslo_vmware.api [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.402912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-22fe7669-eef6-4447-8c04-980cf583aae0 tempest-ListServerFiltersTestJSON-1756963749 tempest-ListServerFiltersTestJSON-1756963749-project-member] Lock "eccde807-664e-4488-bec8-2d6919c2fc91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.348s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.441577] env[62235]: INFO nova.scheduler.client.report [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted allocation for migration 58fa497c-b13d-4311-b06e-5c36066eb749 [ 997.473834] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525474e0-a874-7786-0475-061edf3f99c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010343} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.477090] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eba04942-a86e-4828-9663-6fb51bdc56b3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.483028] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 997.483028] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b8c5cf-2b9f-d595-fafe-6db5b7e34dfc" [ 997.483028] env[62235]: _type = "Task" [ 997.483028] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.490963] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b8c5cf-2b9f-d595-fafe-6db5b7e34dfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.526664] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c30ccbf-2416-4ead-9b49-442570586f20 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.533779] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0427968f-3679-4565-a612-b7f209232885 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.563509] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c8856f-cb79-41ac-9610-1d45a0e5cb48 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.570611] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5c0ab9-d662-4096-912f-0596c577de66 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.583594] env[62235]: DEBUG nova.compute.provider_tree [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.799777] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272094, 'name': Rename_Task, 'duration_secs': 0.1571} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.800093] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.800349] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70d57ab2-3b52-4c37-8f81-a61efd0faf48 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.806859] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 997.806859] env[62235]: value = "task-1272097" [ 997.806859] env[62235]: _type = "Task" [ 997.806859] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.814530] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.870784] env[62235]: DEBUG oslo_vmware.api [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146133} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.871059] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.871255] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.871442] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.871627] env[62235]: INFO nova.compute.manager [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 997.871999] env[62235]: DEBUG oslo.service.loopingcall [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.872239] env[62235]: DEBUG nova.compute.manager [-] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 997.872335] env[62235]: DEBUG nova.network.neutron [-] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 997.910143] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.910396] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.910590] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleting the datastore file [datastore1] 0071c426-0ed5-4e75-bf7f-cbf6221cda07 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.912735] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d37171f-933e-4a37-bfd0-b0bcc07e3134 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.919133] env[62235]: DEBUG oslo_vmware.api [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 997.919133] env[62235]: value = "task-1272098" [ 997.919133] env[62235]: _type = "Task" [ 997.919133] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.928178] env[62235]: DEBUG oslo_vmware.api [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272098, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.947373] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b8dd0d99-a42f-4eeb-9cbf-3e403b68b277 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.413s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.994360] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b8c5cf-2b9f-d595-fafe-6db5b7e34dfc, 'name': SearchDatastore_Task, 'duration_secs': 0.011376} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.994421] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.994923] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] e98e81bd-0bde-4b23-b380-18544512dc20/e98e81bd-0bde-4b23-b380-18544512dc20.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.996320] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e89e543a-80b8-4361-a540-d93b5238627f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.005971] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 998.005971] env[62235]: value = "task-1272099" [ 998.005971] env[62235]: _type = "Task" [ 998.005971] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.016087] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272099, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.087573] env[62235]: DEBUG nova.scheduler.client.report [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.207458] env[62235]: DEBUG nova.compute.manager [req-de42387e-5f20-4f77-9756-c240418dfd93 req-a5f81be4-1fd1-48be-8c02-613fe1268289 service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Received event network-vif-deleted-3e813a1d-b62c-400a-a376-ec445a16c4ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.207802] env[62235]: INFO nova.compute.manager [req-de42387e-5f20-4f77-9756-c240418dfd93 req-a5f81be4-1fd1-48be-8c02-613fe1268289 service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Neutron deleted interface 3e813a1d-b62c-400a-a376-ec445a16c4ca; detaching it from the instance and deleting it from the info cache [ 998.207998] env[62235]: DEBUG nova.network.neutron [req-de42387e-5f20-4f77-9756-c240418dfd93 req-a5f81be4-1fd1-48be-8c02-613fe1268289 service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.319110] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272097, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.429991] env[62235]: DEBUG oslo_vmware.api [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272098, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168947} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.430329] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.430568] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 998.430797] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.430970] env[62235]: INFO nova.compute.manager [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Took 1.43 seconds to destroy the instance on the hypervisor. [ 998.431290] env[62235]: DEBUG oslo.service.loopingcall [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.431506] env[62235]: DEBUG nova.compute.manager [-] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 998.432224] env[62235]: DEBUG nova.network.neutron [-] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 998.516158] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272099, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453216} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.516459] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] e98e81bd-0bde-4b23-b380-18544512dc20/e98e81bd-0bde-4b23-b380-18544512dc20.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 998.516672] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.516932] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-036e8aa4-4217-4c95-a90a-12b2e64ffead {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.523458] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 998.523458] env[62235]: value = "task-1272100" [ 998.523458] env[62235]: _type = "Task" [ 998.523458] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.530871] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272100, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.593342] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.595900] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.477s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.596166] env[62235]: DEBUG nova.objects.instance [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'resources' on Instance uuid a3336b20-5737-4e5a-94a1-19fec140aa60 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.627474] env[62235]: INFO nova.scheduler.client.report [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Deleted allocations for instance 8a6962ba-4406-4ac9-8447-b839903781fe [ 998.650648] env[62235]: DEBUG nova.network.neutron [-] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.710742] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8c0cac0-c78c-45cc-b0d5-084aaf55c3f3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.720595] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5dde11-6313-4809-84f5-8ae38669967f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.745088] env[62235]: DEBUG nova.compute.manager [req-de42387e-5f20-4f77-9756-c240418dfd93 req-a5f81be4-1fd1-48be-8c02-613fe1268289 service nova] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Detach interface failed, port_id=3e813a1d-b62c-400a-a376-ec445a16c4ca, reason: Instance 550945fe-52be-408d-9f42-f5cd5e9c489f could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 998.816968] env[62235]: DEBUG oslo_vmware.api [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272097, 'name': PowerOnVM_Task, 'duration_secs': 0.551103} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.817286] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.817541] env[62235]: INFO nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Took 7.95 seconds to spawn the instance on the hypervisor. [ 998.817801] env[62235]: DEBUG nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.818577] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f196fba-7403-454b-93d0-974e9f721cb6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.033712] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272100, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.170578} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.034029] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.034836] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd1227a-d602-4ac0-93e6-1e6231ed7be2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.048245] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.048499] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.048704] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.048890] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.049103] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.059213] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] e98e81bd-0bde-4b23-b380-18544512dc20/e98e81bd-0bde-4b23-b380-18544512dc20.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.059772] env[62235]: INFO nova.compute.manager [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Terminating instance [ 999.061471] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68c143f2-24fe-4f1b-88e5-3f20792def6a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.075767] env[62235]: DEBUG nova.compute.manager [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.075959] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.076711] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8725a1b4-0ff9-4093-8d2e-a07e62c2f4bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.084048] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.085200] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f62c597e-abf4-49ac-acec-1c28a224ec5f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.086602] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 999.086602] env[62235]: value = "task-1272101" [ 999.086602] env[62235]: _type = "Task" [ 999.086602] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.090840] env[62235]: DEBUG oslo_vmware.api [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 999.090840] env[62235]: value = "task-1272102" [ 999.090840] env[62235]: _type = "Task" [ 999.090840] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.096390] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272101, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.103354] env[62235]: DEBUG oslo_vmware.api [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.136413] env[62235]: DEBUG oslo_concurrency.lockutils [None req-aca44cba-d8c8-4dbf-a2ee-680dc2103262 tempest-ServerDiskConfigTestJSON-1702513981 tempest-ServerDiskConfigTestJSON-1702513981-project-member] Lock "8a6962ba-4406-4ac9-8447-b839903781fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.819s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.151663] env[62235]: DEBUG nova.network.neutron [-] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.156147] env[62235]: INFO nova.compute.manager [-] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Took 1.28 seconds to deallocate network for instance. [ 999.307817] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024543dc-f054-4618-a523-169b9b1f0c61 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.317103] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74929b94-b920-421c-a3b2-2e6069113cec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.353039] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5f532c-a72d-46e8-a18f-d71dcaff20b4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.358014] env[62235]: INFO nova.compute.manager [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Took 14.24 seconds to build instance. [ 999.363234] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138e8521-f728-45f3-9d6c-6c48933c0a6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.378211] env[62235]: DEBUG nova.compute.provider_tree [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.604665] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272101, 'name': ReconfigVM_Task, 'duration_secs': 0.29127} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.609502] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Reconfigured VM instance instance-00000060 to attach disk [datastore2] e98e81bd-0bde-4b23-b380-18544512dc20/e98e81bd-0bde-4b23-b380-18544512dc20.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.610495] env[62235]: DEBUG oslo_vmware.api [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272102, 'name': PowerOffVM_Task, 'duration_secs': 0.182259} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.610807] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a32d0bea-efff-40a3-b2fb-7279e3b4788f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.613226] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 999.613489] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.613831] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4ba35cb-16a9-4285-8a44-539bd6e4e294 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.621138] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 999.621138] env[62235]: value = "task-1272103" [ 999.621138] env[62235]: _type = "Task" [ 999.621138] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.636110] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272103, 'name': Rename_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.655102] env[62235]: INFO nova.compute.manager [-] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Took 1.22 seconds to deallocate network for instance. [ 999.664356] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.707389] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.708495] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.708495] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleting the datastore file [datastore2] 3ea7db36-ebde-4523-b803-ad121a00269b {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.708495] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb64404a-1d76-4839-8dea-0484e53b7715 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.715135] env[62235]: DEBUG oslo_vmware.api [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 999.715135] env[62235]: value = "task-1272105" [ 999.715135] env[62235]: _type = "Task" [ 999.715135] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.724492] env[62235]: DEBUG oslo_vmware.api [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272105, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.864423] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ca5d49fa-459d-4526-b7e5-440e803f0312 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.758s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.881874] env[62235]: DEBUG nova.scheduler.client.report [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.133373] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272103, 'name': Rename_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.166315] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.191693] env[62235]: DEBUG nova.compute.manager [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.191926] env[62235]: DEBUG nova.compute.manager [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing instance network info cache due to event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.192641] env[62235]: DEBUG oslo_concurrency.lockutils [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.192765] env[62235]: DEBUG oslo_concurrency.lockutils [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.192991] env[62235]: DEBUG nova.network.neutron [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.229960] env[62235]: DEBUG oslo_vmware.api [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272105, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160068} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.229960] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.230578] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.231148] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.234022] env[62235]: INFO nova.compute.manager [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1000.234022] env[62235]: DEBUG oslo.service.loopingcall [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.234022] env[62235]: DEBUG nova.compute.manager [-] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.234022] env[62235]: DEBUG nova.network.neutron [-] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.240888] env[62235]: DEBUG nova.compute.manager [req-bbbe202d-1f35-4fe1-937a-b5bbed3cbee2 req-961d92bf-8618-4184-bed9-82cd988e6bd5 service nova] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Received event network-vif-deleted-773479e6-82f4-4668-be25-fd24885c78ed {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.391758] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.395628] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.730s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.395628] env[62235]: DEBUG nova.objects.instance [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lazy-loading 'resources' on Instance uuid 550945fe-52be-408d-9f42-f5cd5e9c489f {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.462917] env[62235]: INFO nova.scheduler.client.report [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleted allocations for instance a3336b20-5737-4e5a-94a1-19fec140aa60 [ 1000.634486] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272103, 'name': Rename_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.937762] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8fdea78f-580c-4605-b46c-cb7b7d64b9a3 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "a3336b20-5737-4e5a-94a1-19fec140aa60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.838s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.965462] env[62235]: DEBUG nova.network.neutron [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updated VIF entry in instance network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1000.966088] env[62235]: DEBUG nova.network.neutron [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap808e6de1-f7", "ovs_interfaceid": "808e6de1-f745-433c-b5e8-705d1fb7466b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.024312] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108447e5-1332-43b7-9651-a6648ec835c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.033212] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384a173b-1806-481d-adfd-6b449b2043b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.069974] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff22ab9-dee8-41b4-b3b5-e845a0ef14f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.078128] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd27980-6a36-41e5-9865-6d09b262bc67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.093472] env[62235]: DEBUG nova.compute.provider_tree [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.135277] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272103, 'name': Rename_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.424710] env[62235]: DEBUG nova.network.neutron [-] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.468962] env[62235]: DEBUG oslo_concurrency.lockutils [req-a8a35098-aa67-41ea-9912-be058c3aa331 req-3f285e5f-f66b-4ce0-b839-03eb5dada35e service nova] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.597105] env[62235]: DEBUG nova.scheduler.client.report [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.635537] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272103, 'name': Rename_Task, 'duration_secs': 1.890103} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.635659] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.635881] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0444ac8b-2cb2-4fe4-8a25-15d4301c3a14 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.642963] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1001.642963] env[62235]: value = "task-1272106" [ 1001.642963] env[62235]: _type = "Task" [ 1001.642963] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.651852] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272106, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.929341] env[62235]: INFO nova.compute.manager [-] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Took 1.70 seconds to deallocate network for instance. [ 1002.102047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.105185] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.938s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.105185] env[62235]: DEBUG nova.objects.instance [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'resources' on Instance uuid 0071c426-0ed5-4e75-bf7f-cbf6221cda07 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.129026] env[62235]: INFO nova.scheduler.client.report [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Deleted allocations for instance 550945fe-52be-408d-9f42-f5cd5e9c489f [ 1002.137635] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "2068403d-5870-4daa-b4d6-acaf39209488" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.139711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "2068403d-5870-4daa-b4d6-acaf39209488" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.139999] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "2068403d-5870-4daa-b4d6-acaf39209488-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.140230] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "2068403d-5870-4daa-b4d6-acaf39209488-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.140415] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "2068403d-5870-4daa-b4d6-acaf39209488-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.143324] env[62235]: INFO nova.compute.manager [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Terminating instance [ 1002.148448] env[62235]: DEBUG nova.compute.manager [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1002.148593] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.149583] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8ccfb1-d900-4ff9-97fd-4d968a7050d5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.157923] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272106, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.160114] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.160430] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5cebbddc-daa4-40f6-beb0-199d298fe42a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.166050] env[62235]: DEBUG oslo_vmware.api [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 1002.166050] env[62235]: value = "task-1272107" [ 1002.166050] env[62235]: _type = "Task" [ 1002.166050] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.175821] env[62235]: DEBUG oslo_vmware.api [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.271990] env[62235]: DEBUG nova.compute.manager [req-e1692f8f-682c-40a3-a5c4-42cc36e5e31d req-aea58cc8-9018-4d6a-ac1f-e75aa6eabff1 service nova] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Received event network-vif-deleted-eed03567-8c20-4e56-902c-fcd6bb25b56b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.434696] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.637812] env[62235]: DEBUG oslo_concurrency.lockutils [None req-beeb206e-da93-4eac-ae10-7376501cceaa tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "550945fe-52be-408d-9f42-f5cd5e9c489f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.918s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.654184] env[62235]: DEBUG oslo_vmware.api [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272106, 'name': PowerOnVM_Task, 'duration_secs': 0.591798} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.655080] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.655292] env[62235]: INFO nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Took 9.44 seconds to spawn the instance on the hypervisor. [ 1002.655477] env[62235]: DEBUG nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.656261] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d959bf-abae-4164-afab-d9c15bd3b2ef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.680570] env[62235]: DEBUG oslo_vmware.api [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272107, 'name': PowerOffVM_Task, 'duration_secs': 0.430468} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.680835] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.681008] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.683692] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5b88650-d9c6-421c-9d41-8be64d5de8f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.728347] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81a7281-01c7-4814-83e9-86baafed3ffa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.736180] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741b4100-f702-4c91-9481-88010080ec6f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.770579] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62206b18-997f-4a48-818a-dc353d86f708 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.773213] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.773442] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.773630] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleting the datastore file [datastore1] 2068403d-5870-4daa-b4d6-acaf39209488 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.773876] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b08a7830-0227-477c-8ae8-c0ac44ff3793 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.781392] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a47a08f-607e-439e-973d-73d2a331430a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.785150] env[62235]: DEBUG oslo_vmware.api [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for the task: (returnval){ [ 1002.785150] env[62235]: value = "task-1272109" [ 1002.785150] env[62235]: _type = "Task" [ 1002.785150] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.798291] env[62235]: DEBUG nova.compute.provider_tree [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.803943] env[62235]: DEBUG oslo_vmware.api [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.183061] env[62235]: INFO nova.compute.manager [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Took 16.93 seconds to build instance. [ 1003.296436] env[62235]: DEBUG oslo_vmware.api [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Task: {'id': task-1272109, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25614} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.296744] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.296960] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.297187] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.297417] env[62235]: INFO nova.compute.manager [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1003.297669] env[62235]: DEBUG oslo.service.loopingcall [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.297910] env[62235]: DEBUG nova.compute.manager [-] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.297955] env[62235]: DEBUG nova.network.neutron [-] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1003.300490] env[62235]: DEBUG nova.scheduler.client.report [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.526472] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.526898] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.527253] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.527566] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.527869] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.530719] env[62235]: INFO nova.compute.manager [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Terminating instance [ 1003.533221] env[62235]: DEBUG nova.compute.manager [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.533551] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.534823] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdde9ba-0310-4f06-88a5-ce1aa70605b7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.546016] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.546393] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3a66dad-ca37-47e3-9f08-66bd6da767be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.553428] env[62235]: DEBUG oslo_vmware.api [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 1003.553428] env[62235]: value = "task-1272110" [ 1003.553428] env[62235]: _type = "Task" [ 1003.553428] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.565483] env[62235]: DEBUG oslo_vmware.api [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272110, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.686046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ece3e2a9-2d3f-4a86-9789-4cb9762d651a tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.444s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.806745] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.807772] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.373s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.808825] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.832842] env[62235]: INFO nova.scheduler.client.report [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted allocations for instance 0071c426-0ed5-4e75-bf7f-cbf6221cda07 [ 1003.840734] env[62235]: INFO nova.scheduler.client.report [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted allocations for instance 3ea7db36-ebde-4523-b803-ad121a00269b [ 1004.057099] env[62235]: DEBUG nova.compute.manager [req-747c8f2a-e805-49da-aa9f-41ccb9d51489 req-f6e3ae8b-cce9-4ac5-be24-16651246224e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Received event network-vif-deleted-64d54817-98a8-4e2e-b317-957980c86633 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.057666] env[62235]: INFO nova.compute.manager [req-747c8f2a-e805-49da-aa9f-41ccb9d51489 req-f6e3ae8b-cce9-4ac5-be24-16651246224e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Neutron deleted interface 64d54817-98a8-4e2e-b317-957980c86633; detaching it from the instance and deleting it from the info cache [ 1004.057666] env[62235]: DEBUG nova.network.neutron [req-747c8f2a-e805-49da-aa9f-41ccb9d51489 req-f6e3ae8b-cce9-4ac5-be24-16651246224e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.069225] env[62235]: DEBUG oslo_vmware.api [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272110, 'name': PowerOffVM_Task, 'duration_secs': 0.252529} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.069225] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.069225] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1004.069225] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e3e1a37-f067-4773-835b-0847158bae54 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.138080] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.138336] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.138529] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Deleting the datastore file [datastore1] c47ea708-9a8b-44e6-a630-c20cfcd6f004 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.138803] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9c59ca5-9899-4f12-a7b5-3be64fa37365 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.147900] env[62235]: DEBUG oslo_vmware.api [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for the task: (returnval){ [ 1004.147900] env[62235]: value = "task-1272112" [ 1004.147900] env[62235]: _type = "Task" [ 1004.147900] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.156907] env[62235]: DEBUG oslo_vmware.api [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.350010] env[62235]: DEBUG oslo_concurrency.lockutils [None req-02dd1995-bb77-4749-a19b-98f1ea846b1f tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "0071c426-0ed5-4e75-bf7f-cbf6221cda07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.352s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.354019] env[62235]: DEBUG oslo_concurrency.lockutils [None req-38ebe72f-8cfc-4b36-b234-ca1b31c77cbb tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "3ea7db36-ebde-4523-b803-ad121a00269b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.303s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.524259] env[62235]: DEBUG nova.network.neutron [-] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.564190] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-635985f5-806e-4cec-b4b6-2c07ff110b7f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.576105] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6bdf49-7694-4687-b857-d15b26909987 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.608526] env[62235]: DEBUG nova.compute.manager [req-747c8f2a-e805-49da-aa9f-41ccb9d51489 req-f6e3ae8b-cce9-4ac5-be24-16651246224e service nova] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Detach interface failed, port_id=64d54817-98a8-4e2e-b317-957980c86633, reason: Instance 2068403d-5870-4daa-b4d6-acaf39209488 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1004.657543] env[62235]: DEBUG oslo_vmware.api [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Task: {'id': task-1272112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172053} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.657828] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.658556] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.658556] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.658556] env[62235]: INFO nova.compute.manager [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1004.658704] env[62235]: DEBUG oslo.service.loopingcall [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.658916] env[62235]: DEBUG nova.compute.manager [-] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.658978] env[62235]: DEBUG nova.network.neutron [-] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1005.029939] env[62235]: INFO nova.compute.manager [-] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Took 1.73 seconds to deallocate network for instance. [ 1005.274055] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "ef220f90-5581-4a51-b308-12da850782a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.274326] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.288300] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.288741] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.538190] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.538476] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.538712] env[62235]: DEBUG nova.objects.instance [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lazy-loading 'resources' on Instance uuid 2068403d-5870-4daa-b4d6-acaf39209488 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.539760] env[62235]: DEBUG nova.network.neutron [-] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.735653] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "4e973f54-2bcf-4953-89fa-c6d34969eec9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.735958] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.777615] env[62235]: DEBUG nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.791175] env[62235]: DEBUG nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.044609] env[62235]: INFO nova.compute.manager [-] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Took 1.39 seconds to deallocate network for instance. [ 1006.096836] env[62235]: DEBUG nova.compute.manager [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Received event network-changed-1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.097012] env[62235]: DEBUG nova.compute.manager [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Refreshing instance network info cache due to event network-changed-1b0d00c1-f3f2-49d0-b742-9c083a6176fa. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.097334] env[62235]: DEBUG oslo_concurrency.lockutils [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] Acquiring lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.097372] env[62235]: DEBUG oslo_concurrency.lockutils [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] Acquired lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.097539] env[62235]: DEBUG nova.network.neutron [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Refreshing network info cache for port 1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.141951] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddecc15-1722-452e-9cd0-fb41baee8686 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.149730] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123d8526-6041-4d61-be9b-57efd31e6e8f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.180359] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5d2371-dc8b-49e4-95e6-32d11a66bf0d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.187580] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84750a15-5368-4d1e-be1b-e03c9e8a49df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.200326] env[62235]: DEBUG nova.compute.provider_tree [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.238473] env[62235]: DEBUG nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.295168] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.310517] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.551274] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.704536] env[62235]: DEBUG nova.scheduler.client.report [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.757734] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.825912] env[62235]: DEBUG nova.network.neutron [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updated VIF entry in instance network info cache for port 1b0d00c1-f3f2-49d0-b742-9c083a6176fa. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.826341] env[62235]: DEBUG nova.network.neutron [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updating instance_info_cache with network_info: [{"id": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "address": "fa:16:3e:aa:39:fb", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b0d00c1-f3", "ovs_interfaceid": "1b0d00c1-f3f2-49d0-b742-9c083a6176fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.211542] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.213909] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.919s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.215486] env[62235]: INFO nova.compute.claims [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.237026] env[62235]: INFO nova.scheduler.client.report [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Deleted allocations for instance 2068403d-5870-4daa-b4d6-acaf39209488 [ 1007.329506] env[62235]: DEBUG oslo_concurrency.lockutils [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] Releasing lock "refresh_cache-e98e81bd-0bde-4b23-b380-18544512dc20" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.329809] env[62235]: DEBUG nova.compute.manager [req-9257413b-2b37-4366-927e-c1d5ff747151 req-7379fbcc-08cc-4bbc-89ca-75d5d03b946b service nova] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Received event network-vif-deleted-7d53cc57-246b-4fed-a55d-514cd12548d8 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.745369] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1f6a539f-fcce-43ed-9c20-32cdfa1ab769 tempest-AttachInterfacesTestJSON-2095038590 tempest-AttachInterfacesTestJSON-2095038590-project-member] Lock "2068403d-5870-4daa-b4d6-acaf39209488" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.607s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.304043] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a2c9bd-2469-4bfb-8276-969ae4369f51 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.311630] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d29b3d5-9cf0-4b2f-ba62-ff9a021088d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.340555] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f88aa3-442f-41f8-8ff4-fab4bbf47cd8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.347648] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485d8b23-16e2-4641-8006-9ef57420a863 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.361688] env[62235]: DEBUG nova.compute.provider_tree [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.865014] env[62235]: DEBUG nova.scheduler.client.report [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.370231] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.156s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.370746] env[62235]: DEBUG nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1009.373536] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.063s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.375350] env[62235]: INFO nova.compute.claims [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.880337] env[62235]: DEBUG nova.compute.utils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.883793] env[62235]: DEBUG nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.883970] env[62235]: DEBUG nova.network.neutron [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1009.929856] env[62235]: DEBUG nova.policy [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '738ee55c831c48f894efbcef71bc20a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c943a445f67f4021bad11723e0836537', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.251736] env[62235]: DEBUG nova.network.neutron [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Successfully created port: c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.384424] env[62235]: DEBUG nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.508319] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b3b97e-ba24-49f2-893d-a7fc25f19016 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.516168] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08a33be-d096-4bdb-866d-a9cbc91b878f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.547038] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ae4aec-1fce-49f1-ad02-896719c56935 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.554305] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c050dd-55be-4312-ab42-c42e14fd2805 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.570523] env[62235]: DEBUG nova.compute.provider_tree [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.073741] env[62235]: DEBUG nova.scheduler.client.report [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.397120] env[62235]: DEBUG nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.421633] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.421855] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.422030] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.422224] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.422376] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.422531] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.422745] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.422912] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.423096] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.423266] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.423439] env[62235]: DEBUG nova.virt.hardware [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.424359] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93be8e11-f5f4-4215-aa2d-1bc1ca85f92a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.432527] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a474a6d1-b170-4d75-808c-ebb17fb29830 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.579159] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.205s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.579766] env[62235]: DEBUG nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.582966] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.032s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.583677] env[62235]: DEBUG nova.objects.instance [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lazy-loading 'resources' on Instance uuid c47ea708-9a8b-44e6-a630-c20cfcd6f004 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.669440] env[62235]: DEBUG nova.compute.manager [req-36727e7f-5c09-4129-bf83-755c461203cb req-c21404bc-a894-4fa0-9f1a-44354ba7c848 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Received event network-vif-plugged-c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.669674] env[62235]: DEBUG oslo_concurrency.lockutils [req-36727e7f-5c09-4129-bf83-755c461203cb req-c21404bc-a894-4fa0-9f1a-44354ba7c848 service nova] Acquiring lock "ef220f90-5581-4a51-b308-12da850782a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.669957] env[62235]: DEBUG oslo_concurrency.lockutils [req-36727e7f-5c09-4129-bf83-755c461203cb req-c21404bc-a894-4fa0-9f1a-44354ba7c848 service nova] Lock "ef220f90-5581-4a51-b308-12da850782a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.670180] env[62235]: DEBUG oslo_concurrency.lockutils [req-36727e7f-5c09-4129-bf83-755c461203cb req-c21404bc-a894-4fa0-9f1a-44354ba7c848 service nova] Lock "ef220f90-5581-4a51-b308-12da850782a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.670370] env[62235]: DEBUG nova.compute.manager [req-36727e7f-5c09-4129-bf83-755c461203cb req-c21404bc-a894-4fa0-9f1a-44354ba7c848 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] No waiting events found dispatching network-vif-plugged-c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.670542] env[62235]: WARNING nova.compute.manager [req-36727e7f-5c09-4129-bf83-755c461203cb req-c21404bc-a894-4fa0-9f1a-44354ba7c848 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Received unexpected event network-vif-plugged-c81e1539-5f47-4bca-bd85-f3ed740feeeb for instance with vm_state building and task_state spawning. [ 1011.750867] env[62235]: DEBUG nova.network.neutron [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Successfully updated port: c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.086219] env[62235]: DEBUG nova.compute.utils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1012.091217] env[62235]: DEBUG nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1012.091390] env[62235]: DEBUG nova.network.neutron [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1012.129780] env[62235]: DEBUG nova.policy [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b96af7a114d04c0dad3c53f5a2dcd2bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef6547723ccf4be891da50a179dae06f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1012.172988] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ccd7cd-1359-413f-a903-70f74a81b971 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.180912] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb4522f-6a01-4984-b951-558377827a25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.212259] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3248607a-0e28-4c65-9d8a-3419f6a94997 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.219860] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6041d3-2478-40ee-8a62-dab2e395b78d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.234402] env[62235]: DEBUG nova.compute.provider_tree [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.254329] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.254329] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.254611] env[62235]: DEBUG nova.network.neutron [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.392401] env[62235]: DEBUG nova.network.neutron [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Successfully created port: 1fe5b4f9-6e4f-4791-aa0e-26507cf68473 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.593870] env[62235]: DEBUG nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.738413] env[62235]: DEBUG nova.scheduler.client.report [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.783604] env[62235]: DEBUG nova.network.neutron [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1012.911449] env[62235]: DEBUG nova.network.neutron [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.243370] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.660s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.245726] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.488s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.247181] env[62235]: INFO nova.compute.claims [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.271787] env[62235]: INFO nova.scheduler.client.report [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Deleted allocations for instance c47ea708-9a8b-44e6-a630-c20cfcd6f004 [ 1013.414143] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.414504] env[62235]: DEBUG nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Instance network_info: |[{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.414970] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:13:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37fb1918-d178-4e12-93e6-316381e78be4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c81e1539-5f47-4bca-bd85-f3ed740feeeb', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.422267] env[62235]: DEBUG oslo.service.loopingcall [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.422495] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef220f90-5581-4a51-b308-12da850782a7] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.422725] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfc4f834-7a3f-45d3-9265-688fe73121c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.444814] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.444814] env[62235]: value = "task-1272113" [ 1013.444814] env[62235]: _type = "Task" [ 1013.444814] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.452241] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272113, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.602400] env[62235]: DEBUG nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.631101] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.631406] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.631589] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.631785] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.631962] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.632661] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.632661] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.632661] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.632772] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.632938] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.633154] env[62235]: DEBUG nova.virt.hardware [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.634161] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef153d95-23ad-458c-91f5-8216bdbd7d02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.642788] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944c4c87-2af5-480c-ad2c-0b10d566ffda {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.698139] env[62235]: DEBUG nova.compute.manager [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Received event network-changed-c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.698241] env[62235]: DEBUG nova.compute.manager [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Refreshing instance network info cache due to event network-changed-c81e1539-5f47-4bca-bd85-f3ed740feeeb. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1013.698446] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] Acquiring lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.698641] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] Acquired lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.698765] env[62235]: DEBUG nova.network.neutron [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Refreshing network info cache for port c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.779641] env[62235]: DEBUG oslo_concurrency.lockutils [None req-44247789-cdf8-42e4-b9e1-0eff477eb594 tempest-ServerRescueTestJSON-128583543 tempest-ServerRescueTestJSON-128583543-project-member] Lock "c47ea708-9a8b-44e6-a630-c20cfcd6f004" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.253s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.859142] env[62235]: DEBUG nova.network.neutron [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Successfully updated port: 1fe5b4f9-6e4f-4791-aa0e-26507cf68473 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.955699] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272113, 'name': CreateVM_Task, 'duration_secs': 0.296206} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.955699] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef220f90-5581-4a51-b308-12da850782a7] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1013.955699] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.955874] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.956251] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1013.956511] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcaad54e-0ca3-461e-936d-d7ee322b71f9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.960644] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1013.960644] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527a2a51-2ced-3b4f-65d2-3c070ac84cd3" [ 1013.960644] env[62235]: _type = "Task" [ 1013.960644] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.969194] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527a2a51-2ced-3b4f-65d2-3c070ac84cd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.337462] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62949bb-244e-4a62-9f84-da3c70b38bc8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.345225] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffe0647-193c-4aee-92aa-54d5abc5eb43 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.375655] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.375809] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.375964] env[62235]: DEBUG nova.network.neutron [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.380314] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b4572d-c93e-413d-a81f-e788d0e67b13 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.388214] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d922ee3-cfde-400f-81da-d3508119748e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.402197] env[62235]: DEBUG nova.compute.provider_tree [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.416445] env[62235]: DEBUG nova.network.neutron [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updated VIF entry in instance network info cache for port c81e1539-5f47-4bca-bd85-f3ed740feeeb. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.416445] env[62235]: DEBUG nova.network.neutron [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.473029] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527a2a51-2ced-3b4f-65d2-3c070ac84cd3, 'name': SearchDatastore_Task, 'duration_secs': 0.012464} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.473029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.473029] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.473029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.473029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.473029] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.473029] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-739f6439-d2e8-425f-870e-1401439218e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.483019] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.483019] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.483019] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-602bd37c-75e5-41bb-aca1-ba09d76779f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.487046] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1014.487046] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52beb2e1-2bc6-3ee6-299e-baad2c5b389c" [ 1014.487046] env[62235]: _type = "Task" [ 1014.487046] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.495479] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52beb2e1-2bc6-3ee6-299e-baad2c5b389c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.905566] env[62235]: DEBUG nova.scheduler.client.report [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.911353] env[62235]: DEBUG nova.network.neutron [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.918108] env[62235]: DEBUG oslo_concurrency.lockutils [req-f9510241-62fd-4e9f-8135-a1afbb074c0d req-bb2db223-240e-4b87-8db0-5eacbd63711c service nova] Releasing lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.997659] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52beb2e1-2bc6-3ee6-299e-baad2c5b389c, 'name': SearchDatastore_Task, 'duration_secs': 0.008945} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.998444] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4976545-e89b-48cd-9615-80726b85ed56 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.005333] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1015.005333] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5202efcc-313f-51fd-bac4-2bc0849fa1f9" [ 1015.005333] env[62235]: _type = "Task" [ 1015.005333] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.012833] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5202efcc-313f-51fd-bac4-2bc0849fa1f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.112388] env[62235]: DEBUG nova.network.neutron [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance_info_cache with network_info: [{"id": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "address": "fa:16:3e:02:0d:2d", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fe5b4f9-6e", "ovs_interfaceid": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.412181] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.412752] env[62235]: DEBUG nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1015.516342] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5202efcc-313f-51fd-bac4-2bc0849fa1f9, 'name': SearchDatastore_Task, 'duration_secs': 0.01368} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.516621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.516910] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.517198] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2161198-5730-4fd5-99ff-3bd691086ae6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.525136] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1015.525136] env[62235]: value = "task-1272114" [ 1015.525136] env[62235]: _type = "Task" [ 1015.525136] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.532665] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272114, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.615142] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.615456] env[62235]: DEBUG nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Instance network_info: |[{"id": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "address": "fa:16:3e:02:0d:2d", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fe5b4f9-6e", "ovs_interfaceid": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.615891] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:0d:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1fe5b4f9-6e4f-4791-aa0e-26507cf68473', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.623891] env[62235]: DEBUG oslo.service.loopingcall [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.624254] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.624598] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8036cb4-1885-48ba-9a26-2a1b30b1f49a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.649024] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.649024] env[62235]: value = "task-1272115" [ 1015.649024] env[62235]: _type = "Task" [ 1015.649024] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.660102] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272115, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.727015] env[62235]: DEBUG nova.compute.manager [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Received event network-vif-plugged-1fe5b4f9-6e4f-4791-aa0e-26507cf68473 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.727267] env[62235]: DEBUG oslo_concurrency.lockutils [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] Acquiring lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.727476] env[62235]: DEBUG oslo_concurrency.lockutils [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.727665] env[62235]: DEBUG oslo_concurrency.lockutils [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.727805] env[62235]: DEBUG nova.compute.manager [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] No waiting events found dispatching network-vif-plugged-1fe5b4f9-6e4f-4791-aa0e-26507cf68473 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.727971] env[62235]: WARNING nova.compute.manager [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Received unexpected event network-vif-plugged-1fe5b4f9-6e4f-4791-aa0e-26507cf68473 for instance with vm_state building and task_state spawning. [ 1015.728362] env[62235]: DEBUG nova.compute.manager [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Received event network-changed-1fe5b4f9-6e4f-4791-aa0e-26507cf68473 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.728544] env[62235]: DEBUG nova.compute.manager [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Refreshing instance network info cache due to event network-changed-1fe5b4f9-6e4f-4791-aa0e-26507cf68473. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.728736] env[62235]: DEBUG oslo_concurrency.lockutils [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] Acquiring lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.728878] env[62235]: DEBUG oslo_concurrency.lockutils [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] Acquired lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.729059] env[62235]: DEBUG nova.network.neutron [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Refreshing network info cache for port 1fe5b4f9-6e4f-4791-aa0e-26507cf68473 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.917949] env[62235]: DEBUG nova.compute.utils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.919503] env[62235]: DEBUG nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1015.919704] env[62235]: DEBUG nova.network.neutron [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1015.960785] env[62235]: DEBUG nova.policy [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91ef5ad3f73a47409c04c43cb22c5c7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7293923cd5284c49a19e4740404c316c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1016.037347] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272114, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469917} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.037631] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.037848] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.038132] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e882d90-5a4d-4dcb-bb9e-832546bd4d86 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.044814] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1016.044814] env[62235]: value = "task-1272116" [ 1016.044814] env[62235]: _type = "Task" [ 1016.044814] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.053714] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272116, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.158639] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272115, 'name': CreateVM_Task, 'duration_secs': 0.3483} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.158832] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1016.159545] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.159721] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.160064] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1016.160334] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d71c374-dfe9-4bb7-a563-6f59cd3b7c65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.165126] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1016.165126] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52754165-5d29-314f-6eb1-da535891f743" [ 1016.165126] env[62235]: _type = "Task" [ 1016.165126] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.172567] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52754165-5d29-314f-6eb1-da535891f743, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.201029] env[62235]: DEBUG nova.network.neutron [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Successfully created port: 6c2f82d3-b377-42f4-8e79-80cecc1d7894 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1016.419194] env[62235]: DEBUG nova.network.neutron [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updated VIF entry in instance network info cache for port 1fe5b4f9-6e4f-4791-aa0e-26507cf68473. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.420031] env[62235]: DEBUG nova.network.neutron [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance_info_cache with network_info: [{"id": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "address": "fa:16:3e:02:0d:2d", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fe5b4f9-6e", "ovs_interfaceid": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.422492] env[62235]: DEBUG nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1016.554454] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272116, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067479} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.554723] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.555589] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b135b9e-62fb-4d1e-a55b-bdd5a37b14f9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.577076] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.577331] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f620b7ff-35a6-458b-aeb1-f5c3d0e8a965 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.596515] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1016.596515] env[62235]: value = "task-1272117" [ 1016.596515] env[62235]: _type = "Task" [ 1016.596515] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.603897] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.675946] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52754165-5d29-314f-6eb1-da535891f743, 'name': SearchDatastore_Task, 'duration_secs': 0.009007} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.676311] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.676585] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.676897] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.677125] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.677386] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.677678] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8142b5ee-be9c-4262-824a-a505021cac1c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.685638] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.685854] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1016.687032] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b21041e0-b0db-404a-95c2-6645d4a28eec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.694792] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1016.694792] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e8538-db45-63f0-2440-62f9b7919c0e" [ 1016.694792] env[62235]: _type = "Task" [ 1016.694792] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.701883] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e8538-db45-63f0-2440-62f9b7919c0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.922137] env[62235]: DEBUG oslo_concurrency.lockutils [req-1498a616-0c7e-4ffa-9a49-c51422b96604 req-262bdde4-1d8a-4cb2-8630-680971ed9d6f service nova] Releasing lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.105745] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272117, 'name': ReconfigVM_Task, 'duration_secs': 0.274991} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.106066] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfigured VM instance instance-00000061 to attach disk [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.106689] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-47b2df1c-c86c-4f26-8c35-f1c9896cf5de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.113043] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1017.113043] env[62235]: value = "task-1272118" [ 1017.113043] env[62235]: _type = "Task" [ 1017.113043] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.120468] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272118, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.204227] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528e8538-db45-63f0-2440-62f9b7919c0e, 'name': SearchDatastore_Task, 'duration_secs': 0.008264} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.204990] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e1eaac4-1659-403f-a3e4-b0a1121ff816 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.210356] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1017.210356] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527020f7-caef-94a9-9384-ac44cb375375" [ 1017.210356] env[62235]: _type = "Task" [ 1017.210356] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.218729] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527020f7-caef-94a9-9384-ac44cb375375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.430552] env[62235]: DEBUG nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1017.455227] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1017.455489] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1017.455653] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1017.455845] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1017.455996] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1017.456263] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1017.456377] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1017.456536] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1017.456706] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1017.456870] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1017.457054] env[62235]: DEBUG nova.virt.hardware [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.457998] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a5c375-aa1e-4ba1-9a8b-48f4f6c03d80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.466109] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d14f46-661b-45a0-89dc-d6573fcf23b3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.622985] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272118, 'name': Rename_Task, 'duration_secs': 0.124718} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.623301] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.623547] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-983248f2-9448-48f0-a0b3-79bf0de035f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.630267] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1017.630267] env[62235]: value = "task-1272119" [ 1017.630267] env[62235]: _type = "Task" [ 1017.630267] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.638722] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272119, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.650881] env[62235]: DEBUG nova.network.neutron [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Successfully updated port: 6c2f82d3-b377-42f4-8e79-80cecc1d7894 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1017.721028] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527020f7-caef-94a9-9384-ac44cb375375, 'name': SearchDatastore_Task, 'duration_secs': 0.009602} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.721243] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.721510] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] d99925a6-a6ed-4069-95b3-6e89d33745b4/d99925a6-a6ed-4069-95b3-6e89d33745b4.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1017.721792] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-243b42ee-fbec-40ec-a273-f5751817ef61 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.728558] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1017.728558] env[62235]: value = "task-1272120" [ 1017.728558] env[62235]: _type = "Task" [ 1017.728558] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.736300] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272120, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.755200] env[62235]: DEBUG nova.compute.manager [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Received event network-vif-plugged-6c2f82d3-b377-42f4-8e79-80cecc1d7894 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.755200] env[62235]: DEBUG oslo_concurrency.lockutils [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] Acquiring lock "4e973f54-2bcf-4953-89fa-c6d34969eec9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.755200] env[62235]: DEBUG oslo_concurrency.lockutils [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.755302] env[62235]: DEBUG oslo_concurrency.lockutils [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.755427] env[62235]: DEBUG nova.compute.manager [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] No waiting events found dispatching network-vif-plugged-6c2f82d3-b377-42f4-8e79-80cecc1d7894 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1017.755616] env[62235]: WARNING nova.compute.manager [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Received unexpected event network-vif-plugged-6c2f82d3-b377-42f4-8e79-80cecc1d7894 for instance with vm_state building and task_state spawning. [ 1017.755793] env[62235]: DEBUG nova.compute.manager [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Received event network-changed-6c2f82d3-b377-42f4-8e79-80cecc1d7894 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.755956] env[62235]: DEBUG nova.compute.manager [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Refreshing instance network info cache due to event network-changed-6c2f82d3-b377-42f4-8e79-80cecc1d7894. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.756195] env[62235]: DEBUG oslo_concurrency.lockutils [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] Acquiring lock "refresh_cache-4e973f54-2bcf-4953-89fa-c6d34969eec9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.756343] env[62235]: DEBUG oslo_concurrency.lockutils [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] Acquired lock "refresh_cache-4e973f54-2bcf-4953-89fa-c6d34969eec9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.756521] env[62235]: DEBUG nova.network.neutron [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Refreshing network info cache for port 6c2f82d3-b377-42f4-8e79-80cecc1d7894 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.141533] env[62235]: DEBUG oslo_vmware.api [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272119, 'name': PowerOnVM_Task, 'duration_secs': 0.487016} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.142061] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.142061] env[62235]: INFO nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Took 6.74 seconds to spawn the instance on the hypervisor. [ 1018.142284] env[62235]: DEBUG nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.143087] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1c6db9-18c7-41d5-9c9e-1a3b84f95cb6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.153359] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "refresh_cache-4e973f54-2bcf-4953-89fa-c6d34969eec9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.239721] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272120, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446771} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.240133] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] d99925a6-a6ed-4069-95b3-6e89d33745b4/d99925a6-a6ed-4069-95b3-6e89d33745b4.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1018.240407] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1018.240753] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8019223b-8dc1-4a61-94c7-d2649a6ce164 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.247848] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1018.247848] env[62235]: value = "task-1272121" [ 1018.247848] env[62235]: _type = "Task" [ 1018.247848] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.256376] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272121, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.291910] env[62235]: DEBUG nova.network.neutron [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1018.359911] env[62235]: DEBUG nova.network.neutron [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.660124] env[62235]: INFO nova.compute.manager [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Took 12.38 seconds to build instance. [ 1018.757315] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272121, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058795} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.757612] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.758431] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d042fe-f5fe-481d-9341-a97ab77ae997 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.779849] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d99925a6-a6ed-4069-95b3-6e89d33745b4/d99925a6-a6ed-4069-95b3-6e89d33745b4.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.780112] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33345573-72d4-463e-b00a-288735d271a9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.799228] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1018.799228] env[62235]: value = "task-1272122" [ 1018.799228] env[62235]: _type = "Task" [ 1018.799228] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.808679] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272122, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.862569] env[62235]: DEBUG oslo_concurrency.lockutils [req-92feef3f-e96a-47cd-adc1-6d0cf69d93a3 req-7a591a31-7827-40ff-b638-d5830e3a65ba service nova] Releasing lock "refresh_cache-4e973f54-2bcf-4953-89fa-c6d34969eec9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.862953] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquired lock "refresh_cache-4e973f54-2bcf-4953-89fa-c6d34969eec9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.863134] env[62235]: DEBUG nova.network.neutron [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.162685] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4119bc33-6653-47e8-a87c-3b388881a074 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.888s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.310889] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272122, 'name': ReconfigVM_Task, 'duration_secs': 0.26168} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.310889] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d99925a6-a6ed-4069-95b3-6e89d33745b4/d99925a6-a6ed-4069-95b3-6e89d33745b4.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.310889] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65c1c6a6-a376-4a54-b451-3ada01d89f5e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.316787] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1019.316787] env[62235]: value = "task-1272123" [ 1019.316787] env[62235]: _type = "Task" [ 1019.316787] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.323660] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272123, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.393893] env[62235]: DEBUG nova.network.neutron [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1019.504764] env[62235]: DEBUG nova.compute.manager [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Received event network-changed-c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.504929] env[62235]: DEBUG nova.compute.manager [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Refreshing instance network info cache due to event network-changed-c81e1539-5f47-4bca-bd85-f3ed740feeeb. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1019.505358] env[62235]: DEBUG oslo_concurrency.lockutils [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] Acquiring lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.505409] env[62235]: DEBUG oslo_concurrency.lockutils [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] Acquired lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.505566] env[62235]: DEBUG nova.network.neutron [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Refreshing network info cache for port c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.518531] env[62235]: DEBUG nova.network.neutron [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Updating instance_info_cache with network_info: [{"id": "6c2f82d3-b377-42f4-8e79-80cecc1d7894", "address": "fa:16:3e:48:94:47", "network": {"id": "01a1f4b4-7589-4a2e-b157-d02922f76099", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1613972526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7293923cd5284c49a19e4740404c316c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c2f82d3-b3", "ovs_interfaceid": "6c2f82d3-b377-42f4-8e79-80cecc1d7894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.826735] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272123, 'name': Rename_Task, 'duration_secs': 0.130174} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.827566] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.827566] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6326819d-7a02-4449-9114-2a2483aee1cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.833393] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1019.833393] env[62235]: value = "task-1272124" [ 1019.833393] env[62235]: _type = "Task" [ 1019.833393] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.840070] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272124, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.021326] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Releasing lock "refresh_cache-4e973f54-2bcf-4953-89fa-c6d34969eec9" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.021695] env[62235]: DEBUG nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Instance network_info: |[{"id": "6c2f82d3-b377-42f4-8e79-80cecc1d7894", "address": "fa:16:3e:48:94:47", "network": {"id": "01a1f4b4-7589-4a2e-b157-d02922f76099", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1613972526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7293923cd5284c49a19e4740404c316c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c2f82d3-b3", "ovs_interfaceid": "6c2f82d3-b377-42f4-8e79-80cecc1d7894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.022176] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:94:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c2f82d3-b377-42f4-8e79-80cecc1d7894', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.030735] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Creating folder: Project (7293923cd5284c49a19e4740404c316c). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1020.031434] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c20564b-c0e2-4bbb-8719-b680bf89b8f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.043237] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Created folder: Project (7293923cd5284c49a19e4740404c316c) in parent group-v273362. [ 1020.043439] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Creating folder: Instances. Parent ref: group-v273508. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1020.045769] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-182bb821-32a4-4ebb-a82c-f0a514d645ad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.055092] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Created folder: Instances in parent group-v273508. [ 1020.055388] env[62235]: DEBUG oslo.service.loopingcall [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.055600] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.056166] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-547dcf62-3d67-4bd1-ba70-fb2ae671f806 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.074447] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.074447] env[62235]: value = "task-1272127" [ 1020.074447] env[62235]: _type = "Task" [ 1020.074447] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.083869] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272127, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.270457] env[62235]: DEBUG nova.network.neutron [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updated VIF entry in instance network info cache for port c81e1539-5f47-4bca-bd85-f3ed740feeeb. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.270961] env[62235]: DEBUG nova.network.neutron [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.343236] env[62235]: DEBUG oslo_vmware.api [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272124, 'name': PowerOnVM_Task, 'duration_secs': 0.443115} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.343521] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.343726] env[62235]: INFO nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Took 6.74 seconds to spawn the instance on the hypervisor. [ 1020.343925] env[62235]: DEBUG nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.344710] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b49d9ff-e75c-40c5-ae38-f9a1fb3f40f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.584477] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272127, 'name': CreateVM_Task, 'duration_secs': 0.333889} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.584644] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.585377] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.585547] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.585883] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1020.586154] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-433f8539-a563-4279-9ef2-5869bbb0385f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.590453] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1020.590453] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5251bf9a-4501-222f-8f80-bedeeb6318e8" [ 1020.590453] env[62235]: _type = "Task" [ 1020.590453] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.597582] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5251bf9a-4501-222f-8f80-bedeeb6318e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.774455] env[62235]: DEBUG oslo_concurrency.lockutils [req-0926e118-bf91-480f-b9a6-f53468c93930 req-49b9c11c-8831-48d6-99b0-c20e48d03c67 service nova] Releasing lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.861810] env[62235]: INFO nova.compute.manager [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Took 14.56 seconds to build instance. [ 1021.101115] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5251bf9a-4501-222f-8f80-bedeeb6318e8, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.101434] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.101682] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.101923] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.102109] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.102317] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.102580] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ce3f906-9058-41f4-a66c-be3a7a07fd2a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.110559] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.110735] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.111417] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0948b59-293e-4bbc-a5fb-82c7d1b871c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.118369] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1021.118369] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523adfbb-1355-5cdb-4ec1-ecec97d1b0c2" [ 1021.118369] env[62235]: _type = "Task" [ 1021.118369] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.125337] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523adfbb-1355-5cdb-4ec1-ecec97d1b0c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.363846] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51087073-d24b-41f0-b917-cb6ee59e5501 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.075s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.629164] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523adfbb-1355-5cdb-4ec1-ecec97d1b0c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010239} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.629627] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b555bbeb-ac9f-450e-99e7-ad5401be14d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.634574] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1021.634574] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520e6348-ec05-e7d6-0904-05936e0b733d" [ 1021.634574] env[62235]: _type = "Task" [ 1021.634574] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.643367] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520e6348-ec05-e7d6-0904-05936e0b733d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.144440] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520e6348-ec05-e7d6-0904-05936e0b733d, 'name': SearchDatastore_Task, 'duration_secs': 0.009724} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.144713] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.144975] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 4e973f54-2bcf-4953-89fa-c6d34969eec9/4e973f54-2bcf-4953-89fa-c6d34969eec9.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.145254] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4864f87-6402-40a6-96cb-3129686e7c8e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.151500] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1022.151500] env[62235]: value = "task-1272128" [ 1022.151500] env[62235]: _type = "Task" [ 1022.151500] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.158603] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272128, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.574356] env[62235]: DEBUG nova.compute.manager [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Stashing vm_state: active {{(pid=62235) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1022.661973] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272128, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434117} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.662351] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 4e973f54-2bcf-4953-89fa-c6d34969eec9/4e973f54-2bcf-4953-89fa-c6d34969eec9.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1022.662688] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1022.662952] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94b76107-0856-4e7c-8eaf-d8f57c3bf9e0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.670119] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1022.670119] env[62235]: value = "task-1272129" [ 1022.670119] env[62235]: _type = "Task" [ 1022.670119] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.677332] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272129, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.095048] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.095453] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.180054] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062823} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.180349] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.181127] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674b1233-14fb-4d29-b906-6becd398afce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.202048] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 4e973f54-2bcf-4953-89fa-c6d34969eec9/4e973f54-2bcf-4953-89fa-c6d34969eec9.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.202304] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b50e71ec-a11f-4325-b3c7-3b044cc0d574 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.221243] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1023.221243] env[62235]: value = "task-1272130" [ 1023.221243] env[62235]: _type = "Task" [ 1023.221243] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.230659] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272130, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.600807] env[62235]: INFO nova.compute.claims [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.733313] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272130, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.107015] env[62235]: INFO nova.compute.resource_tracker [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating resource usage from migration d1748ebd-bfbb-426c-b321-cef1cf8fb5c0 [ 1024.190255] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6376104-8cfc-4d01-b277-bee61e8b127a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.197374] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c4f704-0d45-4939-a614-548d55d7f849 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.229169] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01074dc6-f065-40a2-8e47-a38a08b3cf24 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.236054] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272130, 'name': ReconfigVM_Task, 'duration_secs': 0.985692} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.238032] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 4e973f54-2bcf-4953-89fa-c6d34969eec9/4e973f54-2bcf-4953-89fa-c6d34969eec9.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.238650] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28454b35-ce84-44f3-8864-09d58abf1d6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.240882] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2964a098-576f-497e-8e63-0390e800d423 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.254923] env[62235]: DEBUG nova.compute.provider_tree [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.257017] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1024.257017] env[62235]: value = "task-1272131" [ 1024.257017] env[62235]: _type = "Task" [ 1024.257017] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.264618] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272131, 'name': Rename_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.762074] env[62235]: DEBUG nova.scheduler.client.report [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.770786] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272131, 'name': Rename_Task, 'duration_secs': 0.125283} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.771096] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1024.771292] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfe0adad-b4a6-4275-9b44-53211d8b8ed1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.777764] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1024.777764] env[62235]: value = "task-1272132" [ 1024.777764] env[62235]: _type = "Task" [ 1024.777764] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.787551] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272132, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.267018] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.171s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.267269] env[62235]: INFO nova.compute.manager [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Migrating [ 1025.291886] env[62235]: DEBUG oslo_vmware.api [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272132, 'name': PowerOnVM_Task, 'duration_secs': 0.481288} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.292563] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1025.292563] env[62235]: INFO nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Took 7.86 seconds to spawn the instance on the hypervisor. [ 1025.292699] env[62235]: DEBUG nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.293672] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2064643b-a852-4164-b9fb-5788e77be90c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.783975] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.784385] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.784385] env[62235]: DEBUG nova.network.neutron [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.809319] env[62235]: INFO nova.compute.manager [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Took 19.07 seconds to build instance. [ 1026.311688] env[62235]: DEBUG oslo_concurrency.lockutils [None req-63bd88d6-a160-4fcf-a9f5-f5b5d456af4e tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.576s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.705185] env[62235]: DEBUG nova.network.neutron [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance_info_cache with network_info: [{"id": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "address": "fa:16:3e:02:0d:2d", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fe5b4f9-6e", "ovs_interfaceid": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.757438] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "4e973f54-2bcf-4953-89fa-c6d34969eec9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.757705] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.757928] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "4e973f54-2bcf-4953-89fa-c6d34969eec9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.758136] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.758316] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.760276] env[62235]: INFO nova.compute.manager [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Terminating instance [ 1026.761919] env[62235]: DEBUG nova.compute.manager [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.762132] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.763011] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29353d2-4ef6-462b-b74b-ce144f021687 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.770720] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1026.770944] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4e1c847-a959-41e5-99b9-76ca3c4fe236 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.777932] env[62235]: DEBUG oslo_vmware.api [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1026.777932] env[62235]: value = "task-1272133" [ 1026.777932] env[62235]: _type = "Task" [ 1026.777932] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.785615] env[62235]: DEBUG oslo_vmware.api [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272133, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.208313] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.287422] env[62235]: DEBUG oslo_vmware.api [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272133, 'name': PowerOffVM_Task, 'duration_secs': 0.375073} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.287677] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1027.287856] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.288111] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae5305e0-cfee-4c1d-aad2-f5ea191afc68 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.347612] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.347852] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.348053] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Deleting the datastore file [datastore2] 4e973f54-2bcf-4953-89fa-c6d34969eec9 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.348329] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4b34d69-6f92-4f89-9fda-9e1dab23432a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.354765] env[62235]: DEBUG oslo_vmware.api [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for the task: (returnval){ [ 1027.354765] env[62235]: value = "task-1272135" [ 1027.354765] env[62235]: _type = "Task" [ 1027.354765] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.363666] env[62235]: DEBUG oslo_vmware.api [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272135, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.864327] env[62235]: DEBUG oslo_vmware.api [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Task: {'id': task-1272135, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134232} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.864684] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.864771] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.864953] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.865149] env[62235]: INFO nova.compute.manager [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1027.865403] env[62235]: DEBUG oslo.service.loopingcall [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.865630] env[62235]: DEBUG nova.compute.manager [-] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.865736] env[62235]: DEBUG nova.network.neutron [-] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1028.101868] env[62235]: DEBUG nova.compute.manager [req-a76beb40-8ca5-4eb2-aca5-512e6c98447c req-2bc8b7db-1ba4-4cc6-b52f-4ce4e9109cd3 service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Received event network-vif-deleted-6c2f82d3-b377-42f4-8e79-80cecc1d7894 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.102014] env[62235]: INFO nova.compute.manager [req-a76beb40-8ca5-4eb2-aca5-512e6c98447c req-2bc8b7db-1ba4-4cc6-b52f-4ce4e9109cd3 service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Neutron deleted interface 6c2f82d3-b377-42f4-8e79-80cecc1d7894; detaching it from the instance and deleting it from the info cache [ 1028.102224] env[62235]: DEBUG nova.network.neutron [req-a76beb40-8ca5-4eb2-aca5-512e6c98447c req-2bc8b7db-1ba4-4cc6-b52f-4ce4e9109cd3 service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.581554] env[62235]: DEBUG nova.network.neutron [-] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.604705] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-327df68d-5c3f-40b5-8749-c9fda54b9412 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.614706] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937ed387-4cfa-42d8-92b1-263abe8b9346 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.639784] env[62235]: DEBUG nova.compute.manager [req-a76beb40-8ca5-4eb2-aca5-512e6c98447c req-2bc8b7db-1ba4-4cc6-b52f-4ce4e9109cd3 service nova] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Detach interface failed, port_id=6c2f82d3-b377-42f4-8e79-80cecc1d7894, reason: Instance 4e973f54-2bcf-4953-89fa-c6d34969eec9 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1028.722668] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443b31ce-44c3-467f-9362-1d9db505303e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.741256] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance 'd99925a6-a6ed-4069-95b3-6e89d33745b4' progress to 0 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1029.084079] env[62235]: INFO nova.compute.manager [-] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Took 1.22 seconds to deallocate network for instance. [ 1029.247678] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.248080] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-674f2fdc-7edf-4ca5-9ca1-fb5145f3c35e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.258685] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1029.258685] env[62235]: value = "task-1272136" [ 1029.258685] env[62235]: _type = "Task" [ 1029.258685] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.267070] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.591632] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.591970] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.592270] env[62235]: DEBUG nova.objects.instance [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lazy-loading 'resources' on Instance uuid 4e973f54-2bcf-4953-89fa-c6d34969eec9 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.768074] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272136, 'name': PowerOffVM_Task, 'duration_secs': 0.19245} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.768358] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.768548] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance 'd99925a6-a6ed-4069-95b3-6e89d33745b4' progress to 17 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1030.173394] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7efd73b-64e9-4eea-91c6-312152a8abe8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.180600] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50348b2a-d7db-49fb-8faa-36ecea275e0c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.209343] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d329b8-0452-46f0-b410-9eb64f57ed94 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.216057] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13eda98b-ebe5-4a55-ab44-caf2cfd13bc9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.228641] env[62235]: DEBUG nova.compute.provider_tree [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.274259] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.274531] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.274704] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.274891] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.275053] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.275210] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.275420] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.275686] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.275825] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.275997] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.276196] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.280971] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-987cfad6-60ed-4c11-8613-94aea4589ba3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.296502] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1030.296502] env[62235]: value = "task-1272137" [ 1030.296502] env[62235]: _type = "Task" [ 1030.296502] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.304348] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272137, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.732085] env[62235]: DEBUG nova.scheduler.client.report [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.806489] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272137, 'name': ReconfigVM_Task, 'duration_secs': 0.166411} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.806791] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance 'd99925a6-a6ed-4069-95b3-6e89d33745b4' progress to 33 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1031.237122] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.256520] env[62235]: INFO nova.scheduler.client.report [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Deleted allocations for instance 4e973f54-2bcf-4953-89fa-c6d34969eec9 [ 1031.313343] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.313621] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.313826] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.314057] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.314221] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.314418] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.314655] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.315090] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.315283] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.315467] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.315646] env[62235]: DEBUG nova.virt.hardware [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.320830] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1031.321348] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d61011f8-9b08-430c-8471-aa859b296d60 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.339698] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1031.339698] env[62235]: value = "task-1272138" [ 1031.339698] env[62235]: _type = "Task" [ 1031.339698] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.347242] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272138, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.764520] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e168031c-54c3-4e15-99e1-39379b253cc4 tempest-InstanceActionsV221TestJSON-370214563 tempest-InstanceActionsV221TestJSON-370214563-project-member] Lock "4e973f54-2bcf-4953-89fa-c6d34969eec9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.007s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.849449] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272138, 'name': ReconfigVM_Task, 'duration_secs': 0.149053} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.849736] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1031.850495] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434b46e2-6120-4b83-8ad0-947373e0a85c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.874252] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d99925a6-a6ed-4069-95b3-6e89d33745b4/d99925a6-a6ed-4069-95b3-6e89d33745b4.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.874505] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-287f58ae-2954-4fa9-83f7-5bd37674aee2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.891274] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1031.891274] env[62235]: value = "task-1272139" [ 1031.891274] env[62235]: _type = "Task" [ 1031.891274] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.898614] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272139, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.400689] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272139, 'name': ReconfigVM_Task, 'duration_secs': 0.254533} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.401922] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d99925a6-a6ed-4069-95b3-6e89d33745b4/d99925a6-a6ed-4069-95b3-6e89d33745b4.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.401922] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance 'd99925a6-a6ed-4069-95b3-6e89d33745b4' progress to 50 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1032.908005] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343413d8-3108-4c8f-8970-ec311b01d6e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.929880] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a02efc-9e7a-41e9-9220-cd32bfa5d7d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.949109] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance 'd99925a6-a6ed-4069-95b3-6e89d33745b4' progress to 67 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1033.509906] env[62235]: DEBUG nova.network.neutron [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Port 1fe5b4f9-6e4f-4791-aa0e-26507cf68473 binding to destination host cpu-1 is already ACTIVE {{(pid=62235) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1034.537960] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.537960] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.537960] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.220383] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "e98e81bd-0bde-4b23-b380-18544512dc20" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.220895] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.581957] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.582233] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.582354] env[62235]: DEBUG nova.network.neutron [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1035.724726] env[62235]: DEBUG nova.compute.utils [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.229602] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.297050] env[62235]: DEBUG nova.network.neutron [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance_info_cache with network_info: [{"id": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "address": "fa:16:3e:02:0d:2d", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fe5b4f9-6e", "ovs_interfaceid": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.800296] env[62235]: DEBUG oslo_concurrency.lockutils [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.239793] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.239902] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.240098] env[62235]: INFO nova.compute.manager [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Shelving [ 1037.296748] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "e98e81bd-0bde-4b23-b380-18544512dc20" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.297017] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.297286] env[62235]: INFO nova.compute.manager [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Attaching volume 55fe9dac-7a61-4f99-a993-2c787fe36f2b to /dev/sdb [ 1037.320805] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d13c012-d78e-4808-a9cc-83c23d6ae40c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.341812] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca12ccc-f1bc-49a6-9f7c-f5aca30941f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.344890] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b591b0e1-5233-4866-b3f1-a03cda137bd1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.351768] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance 'd99925a6-a6ed-4069-95b3-6e89d33745b4' progress to 83 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1037.359017] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3284b77-3a89-42b0-888b-dfa8828d7a85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.370311] env[62235]: DEBUG nova.virt.block_device [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updating existing volume attachment record: 5d66e0de-e404-47e7-a3e0-4481f976e011 {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1037.678869] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "0feb5395-f648-419a-998b-eff941c16464" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.679180] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "0feb5395-f648-419a-998b-eff941c16464" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.748048] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.748353] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5064411-8cd7-4c0d-a518-bd9e7e7d17de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.755838] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1037.755838] env[62235]: value = "task-1272141" [ 1037.755838] env[62235]: _type = "Task" [ 1037.755838] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.764606] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.860056] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.860418] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-263abf4a-bb39-4c0e-87ec-4f9462c3dbd1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.868526] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1037.868526] env[62235]: value = "task-1272142" [ 1037.868526] env[62235]: _type = "Task" [ 1037.868526] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.877371] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272142, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.181569] env[62235]: DEBUG nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1038.265063] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272141, 'name': PowerOffVM_Task, 'duration_secs': 0.20536} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.265348] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.266162] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a100c5d-ef4c-4bb9-b7af-32aa6940beff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.283614] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bf0aa8-14b5-4f95-9607-026aca92f370 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.377585] env[62235]: DEBUG oslo_vmware.api [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272142, 'name': PowerOnVM_Task, 'duration_secs': 0.404021} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.377866] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.378068] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-30c9adbe-f6bf-487a-93ce-7ffda43f3db4 tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance 'd99925a6-a6ed-4069-95b3-6e89d33745b4' progress to 100 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1038.704726] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.705031] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.706875] env[62235]: INFO nova.compute.claims [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.793041] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1038.793395] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-eaec83ad-fadc-4389-bf5e-6e34c27b9e8f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.801471] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1038.801471] env[62235]: value = "task-1272143" [ 1038.801471] env[62235]: _type = "Task" [ 1038.801471] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.809665] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272143, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.311234] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272143, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.797937] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6454368a-51e6-4f37-a5f8-767c881cd66c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.808391] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259d86bf-271a-4391-bf11-8ef9b4086a7d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.814737] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272143, 'name': CreateSnapshot_Task, 'duration_secs': 0.654448} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.815395] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1039.816106] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533534de-d641-46b3-9b4a-7303ed7487a1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.842246] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2707f9d3-ecfc-44d8-93b1-e6991d6585f7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.854052] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75972050-cd1f-41ea-81ae-9c39b0cb3eaa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.867843] env[62235]: DEBUG nova.compute.provider_tree [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.135316] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.135600] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.135797] env[62235]: DEBUG nova.compute.manager [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Going to confirm migration 4 {{(pid=62235) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1040.356964] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1040.357309] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-df200f7d-a898-470a-bdf1-b834e72136ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.367057] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1040.367057] env[62235]: value = "task-1272145" [ 1040.367057] env[62235]: _type = "Task" [ 1040.367057] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.370960] env[62235]: DEBUG nova.scheduler.client.report [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.378979] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.675013] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.675291] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquired lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.675518] env[62235]: DEBUG nova.network.neutron [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1040.675727] env[62235]: DEBUG nova.objects.instance [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'info_cache' on Instance uuid d99925a6-a6ed-4069-95b3-6e89d33745b4 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.878266] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.878729] env[62235]: DEBUG nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1040.881389] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.378379] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.383818] env[62235]: DEBUG nova.compute.utils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.385284] env[62235]: DEBUG nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1041.385933] env[62235]: DEBUG nova.network.neutron [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1041.421917] env[62235]: DEBUG nova.policy [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91b451aa2fe74f6eabee71b0f4f90f0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cd98f2307d24723b3aeb481b535ace9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1041.696146] env[62235]: DEBUG nova.network.neutron [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Successfully created port: 8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1041.878962] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.880463] env[62235]: DEBUG nova.network.neutron [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance_info_cache with network_info: [{"id": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "address": "fa:16:3e:02:0d:2d", "network": {"id": "8bb80785-c793-4c96-89ee-d6ba47a5bb73", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1528772031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef6547723ccf4be891da50a179dae06f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fe5b4f9-6e", "ovs_interfaceid": "1fe5b4f9-6e4f-4791-aa0e-26507cf68473", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.888316] env[62235]: DEBUG nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.917458] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1041.917867] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273511', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'name': 'volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e98e81bd-0bde-4b23-b380-18544512dc20', 'attached_at': '', 'detached_at': '', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'serial': '55fe9dac-7a61-4f99-a993-2c787fe36f2b'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1041.918730] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7ee83c-f84f-4130-b364-b883988f2e98 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.935455] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055582ce-ca91-4664-a249-e4f37f948269 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.961286] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b/volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.961807] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5855dc2b-986d-4613-906c-f42aa0e14f6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.980154] env[62235]: DEBUG oslo_vmware.api [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1041.980154] env[62235]: value = "task-1272146" [ 1041.980154] env[62235]: _type = "Task" [ 1041.980154] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.988094] env[62235]: DEBUG oslo_vmware.api [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272146, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.379655] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.383438] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Releasing lock "refresh_cache-d99925a6-a6ed-4069-95b3-6e89d33745b4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.383730] env[62235]: DEBUG nova.objects.instance [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lazy-loading 'migration_context' on Instance uuid d99925a6-a6ed-4069-95b3-6e89d33745b4 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.490139] env[62235]: DEBUG oslo_vmware.api [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272146, 'name': ReconfigVM_Task, 'duration_secs': 0.354807} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.490445] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Reconfigured VM instance instance-00000060 to attach disk [datastore1] volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b/volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.495396] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff8a99e6-4b9a-4bbc-a613-db8213a2502c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.510500] env[62235]: DEBUG oslo_vmware.api [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1042.510500] env[62235]: value = "task-1272147" [ 1042.510500] env[62235]: _type = "Task" [ 1042.510500] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.521882] env[62235]: DEBUG oslo_vmware.api [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272147, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.883588] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task} progress is 95%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.886750] env[62235]: DEBUG nova.objects.base [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1042.887986] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfcc0c5-aada-4a7a-aafd-275edde03111 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.916879] env[62235]: DEBUG nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.921033] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f03ba77c-a97e-485a-8c8a-4464a3ec9dc1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.928066] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1042.928066] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5249e87f-a13d-8277-26ef-3496b429e494" [ 1042.928066] env[62235]: _type = "Task" [ 1042.928066] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.937865] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5249e87f-a13d-8277-26ef-3496b429e494, 'name': SearchDatastore_Task, 'duration_secs': 0.006519} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.939806] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.940075] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.947289] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.947528] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.947692] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.947904] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.948090] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.948238] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.948457] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.948627] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.948798] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.948968] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.949175] env[62235]: DEBUG nova.virt.hardware [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.950278] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1dddf2a-4a29-46d1-b7f8-de65eba85ff6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.960333] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b292e0-9788-4808-a91f-8fa1b7950270 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.020312] env[62235]: DEBUG oslo_vmware.api [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272147, 'name': ReconfigVM_Task, 'duration_secs': 0.147129} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.020626] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273511', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'name': 'volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e98e81bd-0bde-4b23-b380-18544512dc20', 'attached_at': '', 'detached_at': '', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'serial': '55fe9dac-7a61-4f99-a993-2c787fe36f2b'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1043.098627] env[62235]: DEBUG nova.compute.manager [req-b61b6d2a-519d-4496-a697-de9b26af2219 req-4b0d2863-cba8-4016-b365-7661117e9e0f service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Received event network-vif-plugged-8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.098867] env[62235]: DEBUG oslo_concurrency.lockutils [req-b61b6d2a-519d-4496-a697-de9b26af2219 req-4b0d2863-cba8-4016-b365-7661117e9e0f service nova] Acquiring lock "0feb5395-f648-419a-998b-eff941c16464-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.099104] env[62235]: DEBUG oslo_concurrency.lockutils [req-b61b6d2a-519d-4496-a697-de9b26af2219 req-4b0d2863-cba8-4016-b365-7661117e9e0f service nova] Lock "0feb5395-f648-419a-998b-eff941c16464-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.099310] env[62235]: DEBUG oslo_concurrency.lockutils [req-b61b6d2a-519d-4496-a697-de9b26af2219 req-4b0d2863-cba8-4016-b365-7661117e9e0f service nova] Lock "0feb5395-f648-419a-998b-eff941c16464-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.099480] env[62235]: DEBUG nova.compute.manager [req-b61b6d2a-519d-4496-a697-de9b26af2219 req-4b0d2863-cba8-4016-b365-7661117e9e0f service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] No waiting events found dispatching network-vif-plugged-8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1043.099651] env[62235]: WARNING nova.compute.manager [req-b61b6d2a-519d-4496-a697-de9b26af2219 req-4b0d2863-cba8-4016-b365-7661117e9e0f service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Received unexpected event network-vif-plugged-8e87c7f6-11e9-49d5-8326-57e2140c176f for instance with vm_state building and task_state spawning. [ 1043.181566] env[62235]: DEBUG nova.network.neutron [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Successfully updated port: 8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.380670] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.523100] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da943d1-463c-4125-8785-d94f5f75ee4f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.532293] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad5ff9e-19d6-4942-905d-713b71e8bc04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.563137] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b5a3b0-5d24-4d02-91e8-3f5a08929bc2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.570295] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90372d42-f3ab-4528-b1ef-b4360d8190f5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.585123] env[62235]: DEBUG nova.compute.provider_tree [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.688019] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.688207] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.688679] env[62235]: DEBUG nova.network.neutron [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1043.884068] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272145, 'name': CloneVM_Task, 'duration_secs': 3.03186} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.884444] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Created linked-clone VM from snapshot [ 1043.885431] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05bfca1-eb23-4026-93a2-2c8af0b140a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.894089] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Uploading image ede39f76-6301-4e86-86d4-85e79bd11b32 {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1043.914656] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1043.914656] env[62235]: value = "vm-273513" [ 1043.914656] env[62235]: _type = "VirtualMachine" [ 1043.914656] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1043.914949] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-35fc7212-c801-4219-b315-f3fab1741c65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.921751] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease: (returnval){ [ 1043.921751] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a92ea6-e192-88b2-a1d5-22a7bab34e1f" [ 1043.921751] env[62235]: _type = "HttpNfcLease" [ 1043.921751] env[62235]: } obtained for exporting VM: (result){ [ 1043.921751] env[62235]: value = "vm-273513" [ 1043.921751] env[62235]: _type = "VirtualMachine" [ 1043.921751] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1043.922046] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the lease: (returnval){ [ 1043.922046] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a92ea6-e192-88b2-a1d5-22a7bab34e1f" [ 1043.922046] env[62235]: _type = "HttpNfcLease" [ 1043.922046] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1043.929182] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1043.929182] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a92ea6-e192-88b2-a1d5-22a7bab34e1f" [ 1043.929182] env[62235]: _type = "HttpNfcLease" [ 1043.929182] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1044.067136] env[62235]: DEBUG nova.objects.instance [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'flavor' on Instance uuid e98e81bd-0bde-4b23-b380-18544512dc20 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.087737] env[62235]: DEBUG nova.scheduler.client.report [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.227524] env[62235]: DEBUG nova.network.neutron [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1044.359985] env[62235]: DEBUG nova.network.neutron [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.432021] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1044.432021] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a92ea6-e192-88b2-a1d5-22a7bab34e1f" [ 1044.432021] env[62235]: _type = "HttpNfcLease" [ 1044.432021] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1044.432021] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1044.432021] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a92ea6-e192-88b2-a1d5-22a7bab34e1f" [ 1044.432021] env[62235]: _type = "HttpNfcLease" [ 1044.432021] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1044.432021] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5f4dfd-6a7b-41cb-801d-28db9944aa67 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.439467] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b821ed-47f3-dafa-e190-d7640bd65c71/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1044.439717] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b821ed-47f3-dafa-e190-d7640bd65c71/disk-0.vmdk for reading. {{(pid=62235) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1044.530083] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3596ddd8-1958-437a-a90d-c56fb9a1f194 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.572111] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0db451a7-08f0-4f58-b0e4-ef2bd0308a08 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.275s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.763660] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "e98e81bd-0bde-4b23-b380-18544512dc20" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.764416] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.863809] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.864016] env[62235]: DEBUG nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Instance network_info: |[{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.864488] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:70:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e87c7f6-11e9-49d5-8326-57e2140c176f', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.871871] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating folder: Project (6cd98f2307d24723b3aeb481b535ace9). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.872216] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e62cf769-e040-4507-843b-09dfe882a8af {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.943045] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created folder: Project (6cd98f2307d24723b3aeb481b535ace9) in parent group-v273362. [ 1044.943277] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating folder: Instances. Parent ref: group-v273514. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.943611] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e201edb7-c34f-411e-aafb-47cba1b221e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.953340] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created folder: Instances in parent group-v273514. [ 1044.953577] env[62235]: DEBUG oslo.service.loopingcall [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.953774] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0feb5395-f648-419a-998b-eff941c16464] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.954073] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2c5180f-09cc-4fed-ba42-016e2f312ad8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.975098] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.975098] env[62235]: value = "task-1272151" [ 1044.975098] env[62235]: _type = "Task" [ 1044.975098] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.983982] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272151, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.098042] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.158s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.129201] env[62235]: DEBUG nova.compute.manager [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Received event network-changed-8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.129201] env[62235]: DEBUG nova.compute.manager [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Refreshing instance network info cache due to event network-changed-8e87c7f6-11e9-49d5-8326-57e2140c176f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.129201] env[62235]: DEBUG oslo_concurrency.lockutils [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] Acquiring lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.129445] env[62235]: DEBUG oslo_concurrency.lockutils [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] Acquired lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.129564] env[62235]: DEBUG nova.network.neutron [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Refreshing network info cache for port 8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1045.267042] env[62235]: INFO nova.compute.manager [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Detaching volume 55fe9dac-7a61-4f99-a993-2c787fe36f2b [ 1045.305114] env[62235]: INFO nova.virt.block_device [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Attempting to driver detach volume 55fe9dac-7a61-4f99-a993-2c787fe36f2b from mountpoint /dev/sdb [ 1045.305493] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1045.305788] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273511', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'name': 'volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e98e81bd-0bde-4b23-b380-18544512dc20', 'attached_at': '', 'detached_at': '', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'serial': '55fe9dac-7a61-4f99-a993-2c787fe36f2b'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1045.306771] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f16931-be01-4d16-afc9-0906c32d1b36 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.328719] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f4b53d-e54c-4a3e-9158-3778b39aaf2e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.336098] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637cac14-867b-44ec-a27c-54f9f13ea26b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.356555] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89492362-33b6-408a-ac7b-d4a96753188d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.372824] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] The volume has not been displaced from its original location: [datastore1] volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b/volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1045.379392] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1045.379392] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e8860ae-fa32-41fd-9e73-26891a4e0525 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.398476] env[62235]: DEBUG oslo_vmware.api [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1045.398476] env[62235]: value = "task-1272152" [ 1045.398476] env[62235]: _type = "Task" [ 1045.398476] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.408783] env[62235]: DEBUG oslo_vmware.api [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272152, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.484615] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272151, 'name': CreateVM_Task, 'duration_secs': 0.285712} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.484986] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0feb5395-f648-419a-998b-eff941c16464] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.485873] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.486106] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.486763] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.487087] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31577c67-a666-4f5d-8ce4-9847a17786c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.492117] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1045.492117] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5258f582-6475-80b0-4466-8a94f711f67a" [ 1045.492117] env[62235]: _type = "Task" [ 1045.492117] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.499714] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5258f582-6475-80b0-4466-8a94f711f67a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.667976] env[62235]: INFO nova.scheduler.client.report [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted allocation for migration d1748ebd-bfbb-426c-b321-cef1cf8fb5c0 [ 1045.843712] env[62235]: DEBUG nova.network.neutron [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updated VIF entry in instance network info cache for port 8e87c7f6-11e9-49d5-8326-57e2140c176f. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.844169] env[62235]: DEBUG nova.network.neutron [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.908429] env[62235]: DEBUG oslo_vmware.api [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272152, 'name': ReconfigVM_Task, 'duration_secs': 0.214902} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.908809] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1045.913549] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a19a0f2-e73e-4971-9e19-d7f9f10d00cc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.928402] env[62235]: DEBUG oslo_vmware.api [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1045.928402] env[62235]: value = "task-1272153" [ 1045.928402] env[62235]: _type = "Task" [ 1045.928402] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.936669] env[62235]: DEBUG oslo_vmware.api [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272153, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.003273] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5258f582-6475-80b0-4466-8a94f711f67a, 'name': SearchDatastore_Task, 'duration_secs': 0.013658} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.003730] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.004053] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.004378] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.004598] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.004833] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.005207] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bff4c840-46ce-44b3-b6c6-6e1ac8054997 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.013071] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.013284] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1046.014083] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9e4130a-2312-4572-a6f8-3554d78ff3e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.019241] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1046.019241] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529d860b-f976-1d16-6d5f-d3c314c3068a" [ 1046.019241] env[62235]: _type = "Task" [ 1046.019241] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.027213] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529d860b-f976-1d16-6d5f-d3c314c3068a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.173781] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.038s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.347267] env[62235]: DEBUG oslo_concurrency.lockutils [req-0aaee6c6-c37e-4c46-9efe-d6b1ea7306c4 req-62fd4b98-022a-4e69-8c5d-fa68621ed479 service nova] Releasing lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.439340] env[62235]: DEBUG oslo_vmware.api [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272153, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.529749] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529d860b-f976-1d16-6d5f-d3c314c3068a, 'name': SearchDatastore_Task, 'duration_secs': 0.01541} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.531778] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d06e7df-fea9-4154-a76c-0c789a294640 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.535963] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1046.535963] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a2f0b4-6b4b-0fd1-c801-3a1e2e6ca6dd" [ 1046.535963] env[62235]: _type = "Task" [ 1046.535963] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.543674] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a2f0b4-6b4b-0fd1-c801-3a1e2e6ca6dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.938810] env[62235]: DEBUG oslo_vmware.api [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272153, 'name': ReconfigVM_Task, 'duration_secs': 0.744713} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.939059] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273511', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'name': 'volume-55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e98e81bd-0bde-4b23-b380-18544512dc20', 'attached_at': '', 'detached_at': '', 'volume_id': '55fe9dac-7a61-4f99-a993-2c787fe36f2b', 'serial': '55fe9dac-7a61-4f99-a993-2c787fe36f2b'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1047.046222] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a2f0b4-6b4b-0fd1-c801-3a1e2e6ca6dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009707} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.046514] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.046773] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 0feb5395-f648-419a-998b-eff941c16464/0feb5395-f648-419a-998b-eff941c16464.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1047.047040] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0757146d-456b-4580-9b10-6fa36ffa9970 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.054052] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1047.054052] env[62235]: value = "task-1272154" [ 1047.054052] env[62235]: _type = "Task" [ 1047.054052] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.062689] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.487584] env[62235]: DEBUG nova.objects.instance [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'flavor' on Instance uuid e98e81bd-0bde-4b23-b380-18544512dc20 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.564601] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272154, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44136} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.565028] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 0feb5395-f648-419a-998b-eff941c16464/0feb5395-f648-419a-998b-eff941c16464.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1047.565116] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1047.565360] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef1e4bee-172c-4655-912a-d9ba81dadd21 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.572609] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1047.572609] env[62235]: value = "task-1272155" [ 1047.572609] env[62235]: _type = "Task" [ 1047.572609] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.580525] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272155, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.657945] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.658269] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.658489] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.658679] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.658853] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.661073] env[62235]: INFO nova.compute.manager [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Terminating instance [ 1047.662944] env[62235]: DEBUG nova.compute.manager [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.663169] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1047.664039] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7cb52a-dcd7-4b2e-accb-b12aad3cb0c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.671223] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.671465] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d400ea5-452b-453d-af5a-4f830ac6a9c7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.677170] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1047.677170] env[62235]: value = "task-1272156" [ 1047.677170] env[62235]: _type = "Task" [ 1047.677170] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.685311] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272156, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.082580] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272155, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07068} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.082865] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.083648] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c24c63e-6b72-404b-9dbd-c003d9f74974 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.105970] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 0feb5395-f648-419a-998b-eff941c16464/0feb5395-f648-419a-998b-eff941c16464.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.106307] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-836198ba-53b5-4980-b8f9-3ddcde1c2b55 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.127010] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1048.127010] env[62235]: value = "task-1272157" [ 1048.127010] env[62235]: _type = "Task" [ 1048.127010] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.135605] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272157, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.188670] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272156, 'name': PowerOffVM_Task, 'duration_secs': 0.19784} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.188965] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.189160] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1048.189419] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62aced41-f7ba-493c-83b9-24e87c879360 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.269747] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1048.270114] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1048.270358] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleting the datastore file [datastore2] d99925a6-a6ed-4069-95b3-6e89d33745b4 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1048.270674] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eee9e2f5-3bff-45bb-97e8-108df723d670 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.277955] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for the task: (returnval){ [ 1048.277955] env[62235]: value = "task-1272159" [ 1048.277955] env[62235]: _type = "Task" [ 1048.277955] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.286717] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.495748] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f67aee1f-eb88-44b6-98b0-9f1380933d66 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.732s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.637505] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272157, 'name': ReconfigVM_Task, 'duration_secs': 0.276123} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.637842] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 0feb5395-f648-419a-998b-eff941c16464/0feb5395-f648-419a-998b-eff941c16464.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.638487] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bdc7139-b98c-423a-be52-5080208bc8ec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.645179] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1048.645179] env[62235]: value = "task-1272160" [ 1048.645179] env[62235]: _type = "Task" [ 1048.645179] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.653457] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272160, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.788395] env[62235]: DEBUG oslo_vmware.api [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Task: {'id': task-1272159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16546} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.788670] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.788861] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1048.789058] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1048.789271] env[62235]: INFO nova.compute.manager [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1048.789526] env[62235]: DEBUG oslo.service.loopingcall [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.789726] env[62235]: DEBUG nova.compute.manager [-] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.789821] env[62235]: DEBUG nova.network.neutron [-] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1049.028855] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "e98e81bd-0bde-4b23-b380-18544512dc20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.029152] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.029370] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "e98e81bd-0bde-4b23-b380-18544512dc20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.029554] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.029729] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.031761] env[62235]: INFO nova.compute.manager [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Terminating instance [ 1049.033777] env[62235]: DEBUG nova.compute.manager [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.034027] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.036053] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f48dcd4-2306-46a3-8003-2e356f6d5d10 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.040044] env[62235]: DEBUG nova.compute.manager [req-0485805c-4759-4f78-af43-112f453d7f09 req-e6cadfbb-c75f-479f-a96d-6ba8ed336090 service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Received event network-vif-deleted-1fe5b4f9-6e4f-4791-aa0e-26507cf68473 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.040251] env[62235]: INFO nova.compute.manager [req-0485805c-4759-4f78-af43-112f453d7f09 req-e6cadfbb-c75f-479f-a96d-6ba8ed336090 service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Neutron deleted interface 1fe5b4f9-6e4f-4791-aa0e-26507cf68473; detaching it from the instance and deleting it from the info cache [ 1049.040429] env[62235]: DEBUG nova.network.neutron [req-0485805c-4759-4f78-af43-112f453d7f09 req-e6cadfbb-c75f-479f-a96d-6ba8ed336090 service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.047267] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.047507] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3ef6fe0-5c80-49b9-a237-d384727961fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.054053] env[62235]: DEBUG oslo_vmware.api [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1049.054053] env[62235]: value = "task-1272161" [ 1049.054053] env[62235]: _type = "Task" [ 1049.054053] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.062756] env[62235]: DEBUG oslo_vmware.api [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.155895] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272160, 'name': Rename_Task, 'duration_secs': 0.140597} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.156222] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.156459] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f36e505e-1135-4e56-8060-cf6d9b348a49 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.163859] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1049.163859] env[62235]: value = "task-1272162" [ 1049.163859] env[62235]: _type = "Task" [ 1049.163859] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.172384] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272162, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.513617] env[62235]: DEBUG nova.network.neutron [-] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.543274] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fecde2a6-a522-43dc-b8a9-b5435fd8c188 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.553415] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f38a12-c77a-48f8-bfef-c0957f3c8c05 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.574643] env[62235]: DEBUG oslo_vmware.api [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272161, 'name': PowerOffVM_Task, 'duration_secs': 0.171677} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.574963] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.575170] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.575448] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68f9a10e-9a5b-4b2f-a4ea-9cfbafa6b58b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.586864] env[62235]: DEBUG nova.compute.manager [req-0485805c-4759-4f78-af43-112f453d7f09 req-e6cadfbb-c75f-479f-a96d-6ba8ed336090 service nova] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Detach interface failed, port_id=1fe5b4f9-6e4f-4791-aa0e-26507cf68473, reason: Instance d99925a6-a6ed-4069-95b3-6e89d33745b4 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1049.646980] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.647361] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.647551] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleting the datastore file [datastore2] e98e81bd-0bde-4b23-b380-18544512dc20 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.647857] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9f7d475-c17a-469d-a36f-eb8c9fe593db {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.655706] env[62235]: DEBUG oslo_vmware.api [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1049.655706] env[62235]: value = "task-1272164" [ 1049.655706] env[62235]: _type = "Task" [ 1049.655706] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.664491] env[62235]: DEBUG oslo_vmware.api [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.672858] env[62235]: DEBUG oslo_vmware.api [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272162, 'name': PowerOnVM_Task, 'duration_secs': 0.452243} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.673111] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1049.673335] env[62235]: INFO nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Took 6.76 seconds to spawn the instance on the hypervisor. [ 1049.673587] env[62235]: DEBUG nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.674522] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b208355b-e550-4f37-9192-6d76cd1790a1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.016607] env[62235]: INFO nova.compute.manager [-] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Took 1.23 seconds to deallocate network for instance. [ 1050.165363] env[62235]: DEBUG oslo_vmware.api [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274014} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.165600] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.165861] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.166077] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.166271] env[62235]: INFO nova.compute.manager [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1050.166530] env[62235]: DEBUG oslo.service.loopingcall [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.166724] env[62235]: DEBUG nova.compute.manager [-] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.166820] env[62235]: DEBUG nova.network.neutron [-] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1050.191806] env[62235]: INFO nova.compute.manager [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Took 11.51 seconds to build instance. [ 1050.525186] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.525487] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.525683] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.554410] env[62235]: INFO nova.scheduler.client.report [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Deleted allocations for instance d99925a6-a6ed-4069-95b3-6e89d33745b4 [ 1050.693934] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d371b09-445e-42ca-90c9-3403af4b2c92 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "0feb5395-f648-419a-998b-eff941c16464" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.014s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.061358] env[62235]: DEBUG nova.compute.manager [req-bc338637-8984-4e80-862b-16d43fbe18cb req-0b29551e-99e6-4af2-90b5-a37e9b29a573 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Received event network-vif-deleted-1b0d00c1-f3f2-49d0-b742-9c083a6176fa {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.061570] env[62235]: INFO nova.compute.manager [req-bc338637-8984-4e80-862b-16d43fbe18cb req-0b29551e-99e6-4af2-90b5-a37e9b29a573 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Neutron deleted interface 1b0d00c1-f3f2-49d0-b742-9c083a6176fa; detaching it from the instance and deleting it from the info cache [ 1051.061752] env[62235]: DEBUG nova.network.neutron [req-bc338637-8984-4e80-862b-16d43fbe18cb req-0b29551e-99e6-4af2-90b5-a37e9b29a573 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.068973] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9ac9beb9-bcfb-421b-92cd-ac3d3852667d tempest-DeleteServersTestJSON-336035305 tempest-DeleteServersTestJSON-336035305-project-member] Lock "d99925a6-a6ed-4069-95b3-6e89d33745b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.410s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.132769] env[62235]: DEBUG nova.network.neutron [-] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.286308] env[62235]: DEBUG nova.compute.manager [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Received event network-changed-8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.286511] env[62235]: DEBUG nova.compute.manager [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Refreshing instance network info cache due to event network-changed-8e87c7f6-11e9-49d5-8326-57e2140c176f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1051.286729] env[62235]: DEBUG oslo_concurrency.lockutils [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] Acquiring lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.286883] env[62235]: DEBUG oslo_concurrency.lockutils [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] Acquired lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.287281] env[62235]: DEBUG nova.network.neutron [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Refreshing network info cache for port 8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1051.564208] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c71c2ed7-6cd9-422f-904f-daec790a6ce4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.574794] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221891ff-4943-4996-b2e9-3bc2a82adba8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.604369] env[62235]: DEBUG nova.compute.manager [req-bc338637-8984-4e80-862b-16d43fbe18cb req-0b29551e-99e6-4af2-90b5-a37e9b29a573 service nova] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Detach interface failed, port_id=1b0d00c1-f3f2-49d0-b742-9c083a6176fa, reason: Instance e98e81bd-0bde-4b23-b380-18544512dc20 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1051.635383] env[62235]: INFO nova.compute.manager [-] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Took 1.47 seconds to deallocate network for instance. [ 1052.030392] env[62235]: DEBUG nova.network.neutron [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updated VIF entry in instance network info cache for port 8e87c7f6-11e9-49d5-8326-57e2140c176f. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1052.030850] env[62235]: DEBUG nova.network.neutron [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.110585] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b821ed-47f3-dafa-e190-d7640bd65c71/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1052.111656] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fc8618-55ea-4d7b-90be-4bcae37863cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.118251] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b821ed-47f3-dafa-e190-d7640bd65c71/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1052.118428] env[62235]: ERROR oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b821ed-47f3-dafa-e190-d7640bd65c71/disk-0.vmdk due to incomplete transfer. [ 1052.118676] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ffa07887-e279-4f26-a330-09784c2795c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.125281] env[62235]: DEBUG oslo_vmware.rw_handles [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b821ed-47f3-dafa-e190-d7640bd65c71/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1052.125506] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Uploaded image ede39f76-6301-4e86-86d4-85e79bd11b32 to the Glance image server {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1052.127863] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1052.128129] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b16723b5-9601-48c4-81d2-2de422b9d7c8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.133197] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1052.133197] env[62235]: value = "task-1272166" [ 1052.133197] env[62235]: _type = "Task" [ 1052.133197] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.140893] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272166, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.142994] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.143319] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.143581] env[62235]: DEBUG nova.objects.instance [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'resources' on Instance uuid e98e81bd-0bde-4b23-b380-18544512dc20 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.534409] env[62235]: DEBUG oslo_concurrency.lockutils [req-cb254c12-2d02-42c6-acc2-c3074e609849 req-df6dab2f-5364-4933-8d12-0b1f3009a4e0 service nova] Releasing lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.643629] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272166, 'name': Destroy_Task, 'duration_secs': 0.290665} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.643981] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Destroyed the VM [ 1052.644317] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1052.644548] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1e7dcd3a-25d2-4322-9c90-b45be8be3efd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.656537] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1052.656537] env[62235]: value = "task-1272167" [ 1052.656537] env[62235]: _type = "Task" [ 1052.656537] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.665127] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272167, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.726292] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8a8d0e-45c3-41ad-a92a-a157708461c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.734493] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b2cd6f-3507-4c4b-a80a-4ab431d50089 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.766267] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffcb821-8290-4270-ad5e-4d7d8160181e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.773567] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a792a63f-bdf5-4093-870a-86057fbbe074 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.786672] env[62235]: DEBUG nova.compute.provider_tree [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.166419] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272167, 'name': RemoveSnapshot_Task, 'duration_secs': 0.366415} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.166766] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1053.166945] env[62235]: DEBUG nova.compute.manager [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.167707] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f885461-5548-48a9-867c-2ce75d1c105d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.289325] env[62235]: DEBUG nova.scheduler.client.report [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.679448] env[62235]: INFO nova.compute.manager [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Shelve offloading [ 1053.681130] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.681377] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e01bbfe-6539-4288-9755-8a08629f4aad {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.688767] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1053.688767] env[62235]: value = "task-1272168" [ 1053.688767] env[62235]: _type = "Task" [ 1053.688767] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.696558] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.793796] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.816634] env[62235]: INFO nova.scheduler.client.report [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted allocations for instance e98e81bd-0bde-4b23-b380-18544512dc20 [ 1054.199577] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1054.199861] env[62235]: DEBUG nova.compute.manager [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.200573] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5632b5c4-0db8-4a31-9426-260b1340f613 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.206421] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.206593] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.206777] env[62235]: DEBUG nova.network.neutron [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.325994] env[62235]: DEBUG oslo_concurrency.lockutils [None req-45a613e4-f674-46fa-acab-0f0b137d0e99 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "e98e81bd-0bde-4b23-b380-18544512dc20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.297s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.609481] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.609720] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.609923] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1054.609996] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1054.915603] env[62235]: DEBUG nova.network.neutron [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap808e6de1-f7", "ovs_interfaceid": "808e6de1-f745-433c-b5e8-705d1fb7466b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.114527] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.422753] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.424941] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.425127] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1055.425288] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.969293] env[62235]: DEBUG nova.compute.manager [req-8d36d0b6-0e50-4e20-b6e6-e2836311f8dc req-8a676605-30a0-4951-a08d-a2368d0c4201 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-vif-unplugged-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.969513] env[62235]: DEBUG oslo_concurrency.lockutils [req-8d36d0b6-0e50-4e20-b6e6-e2836311f8dc req-8a676605-30a0-4951-a08d-a2368d0c4201 service nova] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.969709] env[62235]: DEBUG oslo_concurrency.lockutils [req-8d36d0b6-0e50-4e20-b6e6-e2836311f8dc req-8a676605-30a0-4951-a08d-a2368d0c4201 service nova] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.969883] env[62235]: DEBUG oslo_concurrency.lockutils [req-8d36d0b6-0e50-4e20-b6e6-e2836311f8dc req-8a676605-30a0-4951-a08d-a2368d0c4201 service nova] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.970058] env[62235]: DEBUG nova.compute.manager [req-8d36d0b6-0e50-4e20-b6e6-e2836311f8dc req-8a676605-30a0-4951-a08d-a2368d0c4201 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] No waiting events found dispatching network-vif-unplugged-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.970367] env[62235]: WARNING nova.compute.manager [req-8d36d0b6-0e50-4e20-b6e6-e2836311f8dc req-8a676605-30a0-4951-a08d-a2368d0c4201 service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received unexpected event network-vif-unplugged-808e6de1-f745-433c-b5e8-705d1fb7466b for instance with vm_state shelved and task_state shelving_offloading. [ 1056.188115] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.189050] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4614fbc2-3ef2-43b8-907f-7cc4a85095f9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.198508] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.198909] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2365da56-029f-4c47-bf33-ccf3bedff948 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.267987] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.268397] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.268677] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleting the datastore file [datastore1] 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.269010] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b57fffba-d01f-4e73-a204-5d9e847fed2f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.280894] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.281224] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.282506] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1056.282506] env[62235]: value = "task-1272170" [ 1056.282506] env[62235]: _type = "Task" [ 1056.282506] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.291732] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.784269] env[62235]: DEBUG nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.795814] env[62235]: DEBUG oslo_vmware.api [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132264} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.796093] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1056.796286] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1056.796471] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1056.821415] env[62235]: INFO nova.scheduler.client.report [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted allocations for instance 738750ad-931e-45e4-b0c6-8232fe1b9f46 [ 1057.065057] env[62235]: DEBUG nova.compute.manager [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Stashing vm_state: active {{(pid=62235) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1057.288990] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap808e6de1-f7", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.309685] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.309954] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.311484] env[62235]: INFO nova.compute.claims [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.325961] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.586963] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.792270] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.792529] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1057.792718] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.793081] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.793081] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.793207] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.793319] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.793465] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.793626] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1057.793774] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1058.000939] env[62235]: DEBUG nova.compute.manager [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.001328] env[62235]: DEBUG nova.compute.manager [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing instance network info cache due to event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1058.001681] env[62235]: DEBUG oslo_concurrency.lockutils [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.001966] env[62235]: DEBUG oslo_concurrency.lockutils [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.002285] env[62235]: DEBUG nova.network.neutron [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1058.296816] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.398391] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135e2abd-679f-42c4-ad50-844ba0b7cc9d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.406573] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9df910f-38b0-4be7-bfb4-deed5027997d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.437881] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f07a49-e184-493b-8dd1-711e68fc3602 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.445691] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb56342-0284-4b33-a9f5-3dd61b0998fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.463873] env[62235]: DEBUG nova.compute.provider_tree [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.920407] env[62235]: DEBUG nova.network.neutron [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updated VIF entry in instance network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.920783] env[62235]: DEBUG nova.network.neutron [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap808e6de1-f7", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.968555] env[62235]: DEBUG nova.scheduler.client.report [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.423894] env[62235]: DEBUG oslo_concurrency.lockutils [req-76483c00-5c11-4a72-a781-41bd45b748cc req-91ef91d7-0db2-4957-b8e9-19f0c3c9beaa service nova] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.474321] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.475353] env[62235]: DEBUG nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1059.478564] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.153s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.479007] env[62235]: DEBUG nova.objects.instance [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'resources' on Instance uuid 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.632638] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.987350] env[62235]: DEBUG nova.objects.instance [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'numa_topology' on Instance uuid 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.989240] env[62235]: DEBUG nova.compute.utils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.992609] env[62235]: DEBUG nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1059.992609] env[62235]: DEBUG nova.network.neutron [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1060.061234] env[62235]: DEBUG nova.policy [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8a1c9e076304498999fe4c6c1d89c69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a272872fcc55419fb474121b7fefd4c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1060.435890] env[62235]: DEBUG nova.network.neutron [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Successfully created port: c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.492061] env[62235]: DEBUG nova.objects.base [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Object Instance<738750ad-931e-45e4-b0c6-8232fe1b9f46> lazy-loaded attributes: resources,numa_topology {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1060.495035] env[62235]: DEBUG nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.604882] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6dc268-78e1-4d30-a844-5da1b80ba5ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.613060] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38aa0bda-beb2-48e0-a2ec-4c726c60460e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.645502] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb7050f-d5d2-48bf-aa6f-79b327ac3245 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.653720] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbdc6140-fdcc-47f1-b4b5-f7f3a5af8f16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.668018] env[62235]: DEBUG nova.compute.provider_tree [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.888959] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.889223] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.171769] env[62235]: DEBUG nova.scheduler.client.report [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.391486] env[62235]: DEBUG nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1061.506530] env[62235]: DEBUG nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1061.531700] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.531982] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.532168] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.532369] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.532535] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.532697] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.532908] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.533190] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.533374] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.533561] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.533756] env[62235]: DEBUG nova.virt.hardware [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.534641] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516423c2-ae22-43df-af41-c742aba96914 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.542430] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618aa4c0-f001-451c-97d6-367adac614dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.681066] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.202s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.683646] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.097s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.922281] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.990999] env[62235]: DEBUG nova.compute.manager [req-c366fbc7-51a5-49e6-9e21-b7454a88fbea req-42b7e392-e462-4758-9c9a-229201fdcd76 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Received event network-vif-plugged-c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.991259] env[62235]: DEBUG oslo_concurrency.lockutils [req-c366fbc7-51a5-49e6-9e21-b7454a88fbea req-42b7e392-e462-4758-9c9a-229201fdcd76 service nova] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.991510] env[62235]: DEBUG oslo_concurrency.lockutils [req-c366fbc7-51a5-49e6-9e21-b7454a88fbea req-42b7e392-e462-4758-9c9a-229201fdcd76 service nova] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.991640] env[62235]: DEBUG oslo_concurrency.lockutils [req-c366fbc7-51a5-49e6-9e21-b7454a88fbea req-42b7e392-e462-4758-9c9a-229201fdcd76 service nova] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.991811] env[62235]: DEBUG nova.compute.manager [req-c366fbc7-51a5-49e6-9e21-b7454a88fbea req-42b7e392-e462-4758-9c9a-229201fdcd76 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] No waiting events found dispatching network-vif-plugged-c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.991976] env[62235]: WARNING nova.compute.manager [req-c366fbc7-51a5-49e6-9e21-b7454a88fbea req-42b7e392-e462-4758-9c9a-229201fdcd76 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Received unexpected event network-vif-plugged-c973dab3-9201-4fb9-8407-9e0eb19895d4 for instance with vm_state building and task_state spawning. [ 1062.082303] env[62235]: DEBUG nova.network.neutron [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Successfully updated port: c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1062.189953] env[62235]: INFO nova.compute.claims [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.194843] env[62235]: DEBUG oslo_concurrency.lockutils [None req-feead8c3-85b2-4bd9-a8e1-15fc8a6dcc7b tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.955s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.195700] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.563s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.195895] env[62235]: INFO nova.compute.manager [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Unshelving [ 1062.586606] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.586808] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.586989] env[62235]: DEBUG nova.network.neutron [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.697154] env[62235]: INFO nova.compute.resource_tracker [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating resource usage from migration 184e870e-abd3-4f70-b21d-002e64b1d06c [ 1062.776304] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2318661-fd34-4ff1-b400-0aefe53fda6b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.783380] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cadb08-bd18-493d-a414-df6eea69e1a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.812512] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c789a73c-dd74-4849-93e1-15ebf16bcc47 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.819323] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75b3b85-5f37-4f65-9420-884dbd4a4b95 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.833131] env[62235]: DEBUG nova.compute.provider_tree [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.116315] env[62235]: DEBUG nova.network.neutron [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1063.217650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.234174] env[62235]: DEBUG nova.network.neutron [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updating instance_info_cache with network_info: [{"id": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "address": "fa:16:3e:4f:29:99", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc973dab3-92", "ovs_interfaceid": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.336407] env[62235]: DEBUG nova.scheduler.client.report [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.737451] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.737768] env[62235]: DEBUG nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Instance network_info: |[{"id": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "address": "fa:16:3e:4f:29:99", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc973dab3-92", "ovs_interfaceid": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1063.738308] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:29:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c973dab3-9201-4fb9-8407-9e0eb19895d4', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.746058] env[62235]: DEBUG oslo.service.loopingcall [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.746283] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.746508] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d94460d-12f0-4897-9ad9-69c6978d251f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.766903] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.766903] env[62235]: value = "task-1272171" [ 1063.766903] env[62235]: _type = "Task" [ 1063.766903] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.776386] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272171, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.841720] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.158s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.841949] env[62235]: INFO nova.compute.manager [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Migrating [ 1063.848407] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.552s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.848534] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.848666] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1063.848957] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.927s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.850360] env[62235]: INFO nova.compute.claims [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.855554] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bf461b-8843-45c3-bf0e-3debc523980c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.870388] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50701a55-6bc8-4848-adf2-a1b6e8060bf1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.887345] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6348d27a-6bc8-4f7b-beba-1cfb96af778d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.894318] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cedd3f-f11d-4ff8-85c6-139f375e2cc2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.926024] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180630MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1063.926205] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.016493] env[62235]: DEBUG nova.compute.manager [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Received event network-changed-c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.016731] env[62235]: DEBUG nova.compute.manager [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Refreshing instance network info cache due to event network-changed-c973dab3-9201-4fb9-8407-9e0eb19895d4. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1064.016978] env[62235]: DEBUG oslo_concurrency.lockutils [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] Acquiring lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.017169] env[62235]: DEBUG oslo_concurrency.lockutils [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] Acquired lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.017350] env[62235]: DEBUG nova.network.neutron [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Refreshing network info cache for port c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.276980] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272171, 'name': CreateVM_Task, 'duration_secs': 0.298871} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.277347] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.277801] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.277975] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.278320] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1064.278571] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d709553-a950-46f7-89f0-d6df03e5c5ee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.283310] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1064.283310] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bc6250-1e60-9cca-1e53-adf917dd86c5" [ 1064.283310] env[62235]: _type = "Task" [ 1064.283310] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.290763] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bc6250-1e60-9cca-1e53-adf917dd86c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.367057] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.367262] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.367444] env[62235]: DEBUG nova.network.neutron [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1064.694787] env[62235]: DEBUG nova.network.neutron [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updated VIF entry in instance network info cache for port c973dab3-9201-4fb9-8407-9e0eb19895d4. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1064.695149] env[62235]: DEBUG nova.network.neutron [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updating instance_info_cache with network_info: [{"id": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "address": "fa:16:3e:4f:29:99", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc973dab3-92", "ovs_interfaceid": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.793187] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bc6250-1e60-9cca-1e53-adf917dd86c5, 'name': SearchDatastore_Task, 'duration_secs': 0.01012} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.793473] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.793708] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.793942] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.794102] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.794288] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.794569] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1dddb35-ba52-477b-b506-f72efaa7118c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.801870] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.802059] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.802697] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ded8dfe1-669f-4b25-8364-74adc818eaed {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.807445] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1064.807445] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d42219-ece5-0ac6-4e27-826c5fb4c857" [ 1064.807445] env[62235]: _type = "Task" [ 1064.807445] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.814395] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d42219-ece5-0ac6-4e27-826c5fb4c857, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.949390] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d99093-e688-415a-9dd9-262e8a067a55 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.956802] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d478223a-3199-4d53-9852-1664dc27ba77 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.990074] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9beceb0f-8bf2-4ad0-b1cc-14282bb5e60f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.997602] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7829cd63-0cd5-406a-bbf7-7bcaec91203e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.010541] env[62235]: DEBUG nova.compute.provider_tree [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.102365] env[62235]: DEBUG nova.network.neutron [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.198817] env[62235]: DEBUG oslo_concurrency.lockutils [req-7c25dbcd-fe38-43ec-b618-c8b333023eb0 req-f0364cbb-08ff-48d2-9087-96d900d162b3 service nova] Releasing lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.317999] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d42219-ece5-0ac6-4e27-826c5fb4c857, 'name': SearchDatastore_Task, 'duration_secs': 0.00776} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.318832] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc5517df-b50f-4ba7-a33f-923c7112c4fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.324089] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1065.324089] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bc94f3-10a2-6964-c0ce-ae8873962266" [ 1065.324089] env[62235]: _type = "Task" [ 1065.324089] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.331758] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bc94f3-10a2-6964-c0ce-ae8873962266, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.515653] env[62235]: DEBUG nova.scheduler.client.report [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.605806] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.834466] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52bc94f3-10a2-6964-c0ce-ae8873962266, 'name': SearchDatastore_Task, 'duration_secs': 0.009316} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.835200] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.835200] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a993bed1-e346-47e7-9f7f-abdfb6685196/a993bed1-e346-47e7-9f7f-abdfb6685196.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.835376] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4453df5c-1abb-4150-b2c0-ccb7bdca90e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.842271] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1065.842271] env[62235]: value = "task-1272172" [ 1065.842271] env[62235]: _type = "Task" [ 1065.842271] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.849752] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272172, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.020506] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.021087] env[62235]: DEBUG nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1066.024722] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.807s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.024951] env[62235]: DEBUG nova.objects.instance [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'pci_requests' on Instance uuid 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.351916] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272172, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.419518} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.352186] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a993bed1-e346-47e7-9f7f-abdfb6685196/a993bed1-e346-47e7-9f7f-abdfb6685196.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1066.352409] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.352684] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3745f57b-92e1-4e85-850e-a52af4378c44 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.359406] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1066.359406] env[62235]: value = "task-1272173" [ 1066.359406] env[62235]: _type = "Task" [ 1066.359406] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.366676] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272173, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.527840] env[62235]: DEBUG nova.compute.utils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1066.529642] env[62235]: DEBUG nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1066.529642] env[62235]: DEBUG nova.network.neutron [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1066.532917] env[62235]: DEBUG nova.objects.instance [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'numa_topology' on Instance uuid 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.570101] env[62235]: DEBUG nova.policy [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93109c6c348a4975b8c144c37e8c4f82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab1aacbb5a8746769c04f2df5f99d01a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1066.870954] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272173, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067331} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.871264] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.872092] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8c71f4-afd6-4369-b5ff-72d2b7a551e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.876123] env[62235]: DEBUG nova.network.neutron [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Successfully created port: 3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1066.899373] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] a993bed1-e346-47e7-9f7f-abdfb6685196/a993bed1-e346-47e7-9f7f-abdfb6685196.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.899692] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c6e036d-cb2a-4384-8f9c-6be873f03e3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.918898] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1066.918898] env[62235]: value = "task-1272174" [ 1066.918898] env[62235]: _type = "Task" [ 1066.918898] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.926658] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272174, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.035567] env[62235]: DEBUG nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1067.038380] env[62235]: INFO nova.compute.claims [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1067.119306] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e492e0-5233-4eae-9776-475ce00ca073 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.138169] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance 'ef220f90-5581-4a51-b308-12da850782a7' progress to 0 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.428769] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272174, 'name': ReconfigVM_Task, 'duration_secs': 0.267703} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.429110] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Reconfigured VM instance instance-00000065 to attach disk [datastore2] a993bed1-e346-47e7-9f7f-abdfb6685196/a993bed1-e346-47e7-9f7f-abdfb6685196.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.429751] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d737cbc-3526-4e94-9e50-59cc2186b037 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.436839] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1067.436839] env[62235]: value = "task-1272175" [ 1067.436839] env[62235]: _type = "Task" [ 1067.436839] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.446530] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272175, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.644394] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.644674] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc8923f2-3cc6-499a-a612-7db6f72286c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.652049] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1067.652049] env[62235]: value = "task-1272176" [ 1067.652049] env[62235]: _type = "Task" [ 1067.652049] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.659926] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272176, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.946891] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272175, 'name': Rename_Task, 'duration_secs': 0.177751} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.947701] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.947701] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b2e01cd-02a5-49cd-8966-dfad690a442c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.953776] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1067.953776] env[62235]: value = "task-1272177" [ 1067.953776] env[62235]: _type = "Task" [ 1067.953776] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.961258] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.048787] env[62235]: DEBUG nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1068.080187] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.080426] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.081029] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.081029] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.081029] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.081300] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.081300] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.081481] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.081658] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.081828] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.082773] env[62235]: DEBUG nova.virt.hardware [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.082879] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440fadb9-763a-4f18-8fc4-ec9f08e23adb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.093095] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41034e23-4569-4061-a24c-3100b421ead1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.141327] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb87886c-f518-495a-8bd3-0fa18576f12c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.148747] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8992db-35ba-4513-b00e-00c9110b3b79 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.181226] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d09e20-2ac6-4b1d-bf1b-3e02f5888409 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.186292] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272176, 'name': PowerOffVM_Task, 'duration_secs': 0.174944} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.186865] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.187079] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance 'ef220f90-5581-4a51-b308-12da850782a7' progress to 17 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1068.193279] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151a1089-eed9-4d79-8631-4d7dbfe94ab5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.206289] env[62235]: DEBUG nova.compute.provider_tree [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.255114] env[62235]: DEBUG nova.compute.manager [req-f82dd448-64f4-47b0-a40d-56520f013870 req-2ab5d3bf-ed68-4184-9524-01d7ef20dcdd service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received event network-vif-plugged-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.255350] env[62235]: DEBUG oslo_concurrency.lockutils [req-f82dd448-64f4-47b0-a40d-56520f013870 req-2ab5d3bf-ed68-4184-9524-01d7ef20dcdd service nova] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.255562] env[62235]: DEBUG oslo_concurrency.lockutils [req-f82dd448-64f4-47b0-a40d-56520f013870 req-2ab5d3bf-ed68-4184-9524-01d7ef20dcdd service nova] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.255873] env[62235]: DEBUG oslo_concurrency.lockutils [req-f82dd448-64f4-47b0-a40d-56520f013870 req-2ab5d3bf-ed68-4184-9524-01d7ef20dcdd service nova] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.256665] env[62235]: DEBUG nova.compute.manager [req-f82dd448-64f4-47b0-a40d-56520f013870 req-2ab5d3bf-ed68-4184-9524-01d7ef20dcdd service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] No waiting events found dispatching network-vif-plugged-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1068.256865] env[62235]: WARNING nova.compute.manager [req-f82dd448-64f4-47b0-a40d-56520f013870 req-2ab5d3bf-ed68-4184-9524-01d7ef20dcdd service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received unexpected event network-vif-plugged-3b639a75-6787-4039-95d9-4197fe7bff26 for instance with vm_state building and task_state spawning. [ 1068.338204] env[62235]: DEBUG nova.network.neutron [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Successfully updated port: 3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1068.463133] env[62235]: DEBUG oslo_vmware.api [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272177, 'name': PowerOnVM_Task, 'duration_secs': 0.415829} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.463483] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1068.463681] env[62235]: INFO nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Took 6.96 seconds to spawn the instance on the hypervisor. [ 1068.463894] env[62235]: DEBUG nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.464736] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e5ad54-1ed4-44db-b3f2-2ee37e1c6b06 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.696054] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.696230] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.696341] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.696550] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.696711] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.696881] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.697121] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.697303] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.697497] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.697685] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.697882] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.705469] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3681d94-f73a-4e1b-8789-a5ffeba24fb6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.719456] env[62235]: DEBUG nova.scheduler.client.report [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.729468] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1068.729468] env[62235]: value = "task-1272178" [ 1068.729468] env[62235]: _type = "Task" [ 1068.729468] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.740706] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272178, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.842037] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.842037] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.842037] env[62235]: DEBUG nova.network.neutron [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1068.980826] env[62235]: INFO nova.compute.manager [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Took 11.69 seconds to build instance. [ 1069.225211] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.200s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.227445] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.301s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.239334] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272178, 'name': ReconfigVM_Task, 'duration_secs': 0.21464} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.239645] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance 'ef220f90-5581-4a51-b308-12da850782a7' progress to 33 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1069.257457] env[62235]: INFO nova.network.neutron [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating port 808e6de1-f745-433c-b5e8-705d1fb7466b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1069.384502] env[62235]: DEBUG nova.network.neutron [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1069.482908] env[62235]: DEBUG oslo_concurrency.lockutils [None req-62c6da7f-d518-4020-a663-7da60d351de7 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.201s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.560324] env[62235]: DEBUG nova.compute.manager [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Received event network-changed-c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.560324] env[62235]: DEBUG nova.compute.manager [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Refreshing instance network info cache due to event network-changed-c973dab3-9201-4fb9-8407-9e0eb19895d4. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.560324] env[62235]: DEBUG oslo_concurrency.lockutils [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] Acquiring lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.560324] env[62235]: DEBUG oslo_concurrency.lockutils [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] Acquired lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.560324] env[62235]: DEBUG nova.network.neutron [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Refreshing network info cache for port c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1069.580979] env[62235]: DEBUG nova.network.neutron [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b639a75-67", "ovs_interfaceid": "3b639a75-6787-4039-95d9-4197fe7bff26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.745759] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.746072] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.746248] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.746467] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.746626] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.746780] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.746990] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.747336] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.747558] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.747783] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.747999] env[62235]: DEBUG nova.virt.hardware [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.753275] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1069.754240] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d288908-cc03-4074-8f78-087e7cc32b56 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.774318] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1069.774318] env[62235]: value = "task-1272179" [ 1069.774318] env[62235]: _type = "Task" [ 1069.774318] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.784169] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272179, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.083745] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.084121] env[62235]: DEBUG nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Instance network_info: |[{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b639a75-67", "ovs_interfaceid": "3b639a75-6787-4039-95d9-4197fe7bff26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1070.084559] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:3c:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73eeba7c-29e1-4fdf-82b3-d62e63e86051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b639a75-6787-4039-95d9-4197fe7bff26', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.092647] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating folder: Project (ab1aacbb5a8746769c04f2df5f99d01a). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1070.092949] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75b3b52d-fd89-485d-b3c6-a612268f3b91 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.104071] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Created folder: Project (ab1aacbb5a8746769c04f2df5f99d01a) in parent group-v273362. [ 1070.104279] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating folder: Instances. Parent ref: group-v273518. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1070.104525] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-923fb758-877c-408d-8036-8d8cc5408612 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.113843] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Created folder: Instances in parent group-v273518. [ 1070.114095] env[62235]: DEBUG oslo.service.loopingcall [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.114317] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1070.114516] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52c34379-5f2f-4939-a3c2-842f3511969d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.133764] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.133764] env[62235]: value = "task-1272182" [ 1070.133764] env[62235]: _type = "Task" [ 1070.133764] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.143365] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272182, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.235753] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Applying migration context for instance ef220f90-5581-4a51-b308-12da850782a7 as it has an incoming, in-progress migration 184e870e-abd3-4f70-b21d-002e64b1d06c. Migration status is migrating {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1070.236664] env[62235]: INFO nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating resource usage from migration 184e870e-abd3-4f70-b21d-002e64b1d06c [ 1070.256968] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0feb5395-f648-419a-998b-eff941c16464 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1070.257135] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a993bed1-e346-47e7-9f7f-abdfb6685196 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1070.257265] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Migration 184e870e-abd3-4f70-b21d-002e64b1d06c is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1070.257385] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ef220f90-5581-4a51-b308-12da850782a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1070.257503] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 738750ad-931e-45e4-b0c6-8232fe1b9f46 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1070.257618] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 5e52ce21-04a1-41db-b62e-2932bf91538e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1070.257796] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1070.257927] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1070.283637] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272179, 'name': ReconfigVM_Task, 'duration_secs': 0.15659} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.284689] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1070.285871] env[62235]: DEBUG nova.compute.manager [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received event network-changed-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.286114] env[62235]: DEBUG nova.compute.manager [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Refreshing instance network info cache due to event network-changed-3b639a75-6787-4039-95d9-4197fe7bff26. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1070.286252] env[62235]: DEBUG oslo_concurrency.lockutils [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] Acquiring lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.286446] env[62235]: DEBUG oslo_concurrency.lockutils [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] Acquired lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.286561] env[62235]: DEBUG nova.network.neutron [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Refreshing network info cache for port 3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1070.290922] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ca1971-6646-433e-a2ba-84e46f16740d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.294135] env[62235]: DEBUG nova.network.neutron [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updated VIF entry in instance network info cache for port c973dab3-9201-4fb9-8407-9e0eb19895d4. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1070.294462] env[62235]: DEBUG nova.network.neutron [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updating instance_info_cache with network_info: [{"id": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "address": "fa:16:3e:4f:29:99", "network": {"id": "43d25177-0d17-4675-bcb6-5fc60318624f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1175967333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a272872fcc55419fb474121b7fefd4c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc973dab3-92", "ovs_interfaceid": "c973dab3-9201-4fb9-8407-9e0eb19895d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.316353] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.319266] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b87e532-9adb-4403-8147-ac1756bbd8e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.336817] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1070.336817] env[62235]: value = "task-1272183" [ 1070.336817] env[62235]: _type = "Task" [ 1070.336817] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.348415] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272183, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.379890] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cb548c-77f8-46af-8355-8f3df670506d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.387565] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827c23d5-9580-4db2-a130-d7c798cffb6c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.417460] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730d5335-0ab0-4114-94a1-f1cbd4dfefe8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.425513] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af772595-6a98-4f66-bda6-bb8194b224c7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.438611] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.643989] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272182, 'name': CreateVM_Task, 'duration_secs': 0.318966} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.644278] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1070.644929] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.645114] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.645467] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.645725] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8168289c-cb27-468a-8772-6f614b81e87f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.650027] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1070.650027] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521798f2-da38-6cd1-75b7-49a0adb6f8fa" [ 1070.650027] env[62235]: _type = "Task" [ 1070.650027] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.657622] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521798f2-da38-6cd1-75b7-49a0adb6f8fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.748494] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.748658] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.749046] env[62235]: DEBUG nova.network.neutron [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1070.796549] env[62235]: DEBUG oslo_concurrency.lockutils [req-e12bf5d4-c07d-4dda-b052-9cb08cbe149e req-1945d7ea-51d3-4511-a1b1-0da005382ba1 service nova] Releasing lock "refresh_cache-a993bed1-e346-47e7-9f7f-abdfb6685196" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.846374] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272183, 'name': ReconfigVM_Task, 'duration_secs': 0.26705} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.846701] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfigured VM instance instance-00000061 to attach disk [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.846977] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance 'ef220f90-5581-4a51-b308-12da850782a7' progress to 50 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1070.941572] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.019029] env[62235]: DEBUG nova.network.neutron [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updated VIF entry in instance network info cache for port 3b639a75-6787-4039-95d9-4197fe7bff26. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1071.019411] env[62235]: DEBUG nova.network.neutron [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b639a75-67", "ovs_interfaceid": "3b639a75-6787-4039-95d9-4197fe7bff26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.160418] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521798f2-da38-6cd1-75b7-49a0adb6f8fa, 'name': SearchDatastore_Task, 'duration_secs': 0.012528} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.160748] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.160994] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.161251] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.161407] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.161593] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.161858] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e338e27a-3865-4924-8162-8667927d3f71 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.169232] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.169413] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1071.170118] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-478a302c-a3cb-48b9-972d-27c178ebc895 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.174908] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1071.174908] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52285d24-6a83-7d2b-e324-5b845670c002" [ 1071.174908] env[62235]: _type = "Task" [ 1071.174908] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.182961] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52285d24-6a83-7d2b-e324-5b845670c002, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.354193] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03046d1-f8bb-4b2b-974a-07a2408133c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.375145] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e7387d-bf85-4c95-905c-e32549c1d8a6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.393112] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance 'ef220f90-5581-4a51-b308-12da850782a7' progress to 67 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1071.450485] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1071.450684] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.223s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.450897] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.451047] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Cleaning up deleted instances {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1071.466221] env[62235]: DEBUG nova.network.neutron [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap808e6de1-f7", "ovs_interfaceid": "808e6de1-f745-433c-b5e8-705d1fb7466b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.521861] env[62235]: DEBUG oslo_concurrency.lockutils [req-46341aa0-bcbf-4ada-baa6-12cf39499ec3 req-500f9f94-5c58-41da-9e20-294d7636b5f4 service nova] Releasing lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.685838] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52285d24-6a83-7d2b-e324-5b845670c002, 'name': SearchDatastore_Task, 'duration_secs': 0.008589} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.686622] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63c44e67-319f-4e4e-bc9d-0a172368610a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.691549] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1071.691549] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521ce650-1ec0-0676-d720-bec4675eaa82" [ 1071.691549] env[62235]: _type = "Task" [ 1071.691549] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.698641] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521ce650-1ec0-0676-d720-bec4675eaa82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.952654] env[62235]: DEBUG nova.network.neutron [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Port c81e1539-5f47-4bca-bd85-f3ed740feeeb binding to destination host cpu-1 is already ACTIVE {{(pid=62235) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1071.964378] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] There are 44 instances to clean {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1071.964692] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 4e973f54-2bcf-4953-89fa-c6d34969eec9] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.970196] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.998111] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='2d8f7038b82def311c1e41c502a2cbb3',container_format='bare',created_at=2024-10-12T06:27:01Z,direct_url=,disk_format='vmdk',id=ede39f76-6301-4e86-86d4-85e79bd11b32,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-311769504-shelved',owner='6832a93601f84138aeed5a906f24394b',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-12T06:27:16Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1071.998397] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1071.998562] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1071.998755] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1071.998907] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1071.999070] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1071.999291] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1071.999463] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1071.999686] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1071.999869] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1072.000072] env[62235]: DEBUG nova.virt.hardware [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1072.000934] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc0f980-c1ed-486e-a704-9e1aa1c595a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.011225] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326d240f-3b46-4171-ad2b-265c85306986 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.024757] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:c5:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '808e6de1-f745-433c-b5e8-705d1fb7466b', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.031935] env[62235]: DEBUG oslo.service.loopingcall [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.032195] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1072.032413] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2380c7ac-4a5d-46fa-bab7-658ab64eb2f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.051038] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.051038] env[62235]: value = "task-1272184" [ 1072.051038] env[62235]: _type = "Task" [ 1072.051038] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.058747] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272184, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.202175] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521ce650-1ec0-0676-d720-bec4675eaa82, 'name': SearchDatastore_Task, 'duration_secs': 0.008728} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.202472] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.202743] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1072.203017] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed6c5125-d0e2-4f3a-bcf0-af7831542849 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.210345] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1072.210345] env[62235]: value = "task-1272185" [ 1072.210345] env[62235]: _type = "Task" [ 1072.210345] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.218472] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272185, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.312169] env[62235]: DEBUG nova.compute.manager [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-vif-plugged-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.312539] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.312873] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.313165] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.313512] env[62235]: DEBUG nova.compute.manager [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] No waiting events found dispatching network-vif-plugged-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1072.313764] env[62235]: WARNING nova.compute.manager [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received unexpected event network-vif-plugged-808e6de1-f745-433c-b5e8-705d1fb7466b for instance with vm_state shelved_offloaded and task_state spawning. [ 1072.313961] env[62235]: DEBUG nova.compute.manager [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.314155] env[62235]: DEBUG nova.compute.manager [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing instance network info cache due to event network-changed-808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1072.314360] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] Acquiring lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.314503] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] Acquired lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.314668] env[62235]: DEBUG nova.network.neutron [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Refreshing network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1072.471466] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d99925a6-a6ed-4069-95b3-6e89d33745b4] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.563409] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272184, 'name': CreateVM_Task, 'duration_secs': 0.341911} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.563745] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1072.564605] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.564780] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.565222] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1072.565542] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-641bba55-4b58-4b19-b968-2e7dc6af39ba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.571471] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1072.571471] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e8eeb0-e45e-0d69-b84b-a0b3c0e9f1e9" [ 1072.571471] env[62235]: _type = "Task" [ 1072.571471] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.580492] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e8eeb0-e45e-0d69-b84b-a0b3c0e9f1e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.720026] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272185, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434142} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.720353] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.720498] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.720731] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b55477c-13d3-42a2-ac24-8fb75af89ce7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.726609] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1072.726609] env[62235]: value = "task-1272186" [ 1072.726609] env[62235]: _type = "Task" [ 1072.726609] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.734127] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272186, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.981724] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: e98e81bd-0bde-4b23-b380-18544512dc20] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.991102] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "ef220f90-5581-4a51-b308-12da850782a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.991372] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.991554] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.028037] env[62235]: DEBUG nova.network.neutron [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updated VIF entry in instance network info cache for port 808e6de1-f745-433c-b5e8-705d1fb7466b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1073.028424] env[62235]: DEBUG nova.network.neutron [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [{"id": "808e6de1-f745-433c-b5e8-705d1fb7466b", "address": "fa:16:3e:75:c5:20", "network": {"id": "5222bc82-fb20-41b1-9314-834c623d9067", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1693572532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6832a93601f84138aeed5a906f24394b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap808e6de1-f7", "ovs_interfaceid": "808e6de1-f745-433c-b5e8-705d1fb7466b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.082133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.082133] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Processing image ede39f76-6301-4e86-86d4-85e79bd11b32 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.082133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32/ede39f76-6301-4e86-86d4-85e79bd11b32.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.082402] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32/ede39f76-6301-4e86-86d4-85e79bd11b32.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.082469] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.082707] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6396fbc-16b7-4a66-acec-4c9fc00dbe9b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.090628] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.090822] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1073.091525] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00efd8ad-ddff-4048-b302-6f480a4a00d7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.096564] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1073.096564] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d843ef-439f-dbcb-6222-041f48664f5d" [ 1073.096564] env[62235]: _type = "Task" [ 1073.096564] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.103843] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d843ef-439f-dbcb-6222-041f48664f5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.236027] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056099} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.236273] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.237085] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce9171f-4f38-4e3f-b1f7-2600e6c6b786 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.258462] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.259027] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b4197ee-1792-4c66-9336-3ded4be73774 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.278109] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1073.278109] env[62235]: value = "task-1272187" [ 1073.278109] env[62235]: _type = "Task" [ 1073.278109] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.285755] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272187, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.492222] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0071c426-0ed5-4e75-bf7f-cbf6221cda07] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.531321] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccf74cc7-651f-4e60-a1f4-39470a7aa94c req-32200902-12c5-48be-8f3e-5caa2eb5b7bc service nova] Releasing lock "refresh_cache-738750ad-931e-45e4-b0c6-8232fe1b9f46" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.606330] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Preparing fetch location {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1073.606594] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Fetch image to [datastore2] OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098/OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098.vmdk {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1073.606785] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Downloading stream optimized image ede39f76-6301-4e86-86d4-85e79bd11b32 to [datastore2] OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098/OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098.vmdk on the data store datastore2 as vApp {{(pid=62235) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1073.606962] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Downloading image file data ede39f76-6301-4e86-86d4-85e79bd11b32 to the ESX as VM named 'OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098' {{(pid=62235) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1073.672075] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1073.672075] env[62235]: value = "resgroup-9" [ 1073.672075] env[62235]: _type = "ResourcePool" [ 1073.672075] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1073.672393] env[62235]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8eddb7b5-8712-4974-b7a4-243071b1c586 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.692898] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease: (returnval){ [ 1073.692898] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520264f4-ef5f-122c-8512-b0d5b1e6fc5a" [ 1073.692898] env[62235]: _type = "HttpNfcLease" [ 1073.692898] env[62235]: } obtained for vApp import into resource pool (val){ [ 1073.692898] env[62235]: value = "resgroup-9" [ 1073.692898] env[62235]: _type = "ResourcePool" [ 1073.692898] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1073.693352] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the lease: (returnval){ [ 1073.693352] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520264f4-ef5f-122c-8512-b0d5b1e6fc5a" [ 1073.693352] env[62235]: _type = "HttpNfcLease" [ 1073.693352] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1073.702156] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1073.702156] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520264f4-ef5f-122c-8512-b0d5b1e6fc5a" [ 1073.702156] env[62235]: _type = "HttpNfcLease" [ 1073.702156] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1073.787777] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.995756] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8a6962ba-4406-4ac9-8447-b839903781fe] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.029274] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.029484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.029657] env[62235]: DEBUG nova.network.neutron [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1074.201174] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1074.201174] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520264f4-ef5f-122c-8512-b0d5b1e6fc5a" [ 1074.201174] env[62235]: _type = "HttpNfcLease" [ 1074.201174] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1074.288604] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272187, 'name': ReconfigVM_Task, 'duration_secs': 0.791547} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.288894] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.289518] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5bf38a8f-50f9-4e50-a428-eb11ff77ff9b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.296125] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1074.296125] env[62235]: value = "task-1272189" [ 1074.296125] env[62235]: _type = "Task" [ 1074.296125] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.303643] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272189, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.501800] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 550945fe-52be-408d-9f42-f5cd5e9c489f] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.703640] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1074.703640] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520264f4-ef5f-122c-8512-b0d5b1e6fc5a" [ 1074.703640] env[62235]: _type = "HttpNfcLease" [ 1074.703640] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1074.703938] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1074.703938] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]520264f4-ef5f-122c-8512-b0d5b1e6fc5a" [ 1074.703938] env[62235]: _type = "HttpNfcLease" [ 1074.703938] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1074.704682] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fd8485-1a40-47ed-9f1d-8ad381ece476 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.711579] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cbe0c-83c5-400b-85d1-243662df30c7/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1074.711766] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cbe0c-83c5-400b-85d1-243662df30c7/disk-0.vmdk. {{(pid=62235) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1074.768758] env[62235]: DEBUG nova.network.neutron [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.774689] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6e40a3f6-d5f2-42a7-be68-7f11378e72aa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.805186] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272189, 'name': Rename_Task, 'duration_secs': 0.123877} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.805532] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.805693] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1bee77b8-0805-4fbb-8420-1eb3624eb517 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.811902] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1074.811902] env[62235]: value = "task-1272190" [ 1074.811902] env[62235]: _type = "Task" [ 1074.811902] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.821948] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272190, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.005740] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 6f803e24-2b2b-4a15-8172-29acde2154f8] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.271293] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.324349] env[62235]: DEBUG oslo_vmware.api [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272190, 'name': PowerOnVM_Task, 'duration_secs': 0.459079} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.324713] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1075.324932] env[62235]: INFO nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Took 7.28 seconds to spawn the instance on the hypervisor. [ 1075.325135] env[62235]: DEBUG nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.328094] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0441c7bc-2e5a-4264-8e5f-675e98ac256a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.508534] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: c47ea708-9a8b-44e6-a630-c20cfcd6f004] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.804716] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bda22d2-daf7-4c13-92e7-f26d41bb968d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.831530] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d174895-59a3-48f4-97f6-645f045ff0a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.839068] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance 'ef220f90-5581-4a51-b308-12da850782a7' progress to 83 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1075.852924] env[62235]: INFO nova.compute.manager [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Took 13.95 seconds to build instance. [ 1075.887137] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Completed reading data from the image iterator. {{(pid=62235) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1075.887409] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cbe0c-83c5-400b-85d1-243662df30c7/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1075.888877] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385c55cd-d3bf-4ce0-bdec-a8ef8ed909c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.895468] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cbe0c-83c5-400b-85d1-243662df30c7/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1075.895646] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cbe0c-83c5-400b-85d1-243662df30c7/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1075.895891] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-c1ae2a4f-358d-447b-b8f2-6459c77d5333 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.014394] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 79dafae1-e389-4e04-af20-577b8aac06a6] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.126721] env[62235]: DEBUG oslo_vmware.rw_handles [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cbe0c-83c5-400b-85d1-243662df30c7/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1076.126949] env[62235]: INFO nova.virt.vmwareapi.images [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Downloaded image file data ede39f76-6301-4e86-86d4-85e79bd11b32 [ 1076.127844] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff0e678-774e-4bfc-864d-e339832d48f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.143852] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f62c241d-5249-4885-a176-75f98d4ab301 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.197175] env[62235]: INFO nova.virt.vmwareapi.images [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] The imported VM was unregistered [ 1076.199738] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Caching image {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1076.199987] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Creating directory with path [datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1076.200287] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03732034-e6ce-494c-90cd-0d357674c172 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.217815] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Created directory with path [datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1076.218102] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098/OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098.vmdk to [datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32/ede39f76-6301-4e86-86d4-85e79bd11b32.vmdk. {{(pid=62235) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1076.218399] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-6a78bb97-ca72-4fbe-98aa-a47e1c9a1dff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.227148] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1076.227148] env[62235]: value = "task-1272192" [ 1076.227148] env[62235]: _type = "Task" [ 1076.227148] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.235223] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272192, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.350880] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1076.351325] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17b364b1-5837-4c1a-ba94-84b8cf71e51d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.354880] env[62235]: DEBUG oslo_concurrency.lockutils [None req-48cf116f-3f53-44ef-852d-154b797da8b3 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.466s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.358256] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1076.358256] env[62235]: value = "task-1272193" [ 1076.358256] env[62235]: _type = "Task" [ 1076.358256] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.365887] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.518017] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 85e1179a-7afe-497b-8d4f-f3f4bb5ec8bf] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.569199] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.569568] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.740681] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272192, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.869992] env[62235]: DEBUG oslo_vmware.api [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272193, 'name': PowerOnVM_Task, 'duration_secs': 0.42691} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.873078] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.873078] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3996474-85dd-4553-9d27-e4b1797cdb82 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance 'ef220f90-5581-4a51-b308-12da850782a7' progress to 100 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1077.021847] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: eccde807-664e-4488-bec8-2d6919c2fc91] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.071744] env[62235]: DEBUG nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1077.241292] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272192, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.525823] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: a3336b20-5737-4e5a-94a1-19fec140aa60] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.596447] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.596730] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.598375] env[62235]: INFO nova.compute.claims [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.741858] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272192, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.029134] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0ac4202d-4e25-417b-9ab0-6429b73e8e98] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.242838] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272192, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.532147] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 05aae40a-37ae-4bf1-91d0-02a7228fcff2] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.704439] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c24c66-f2a0-469f-8f2e-b5e4d0e28be1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.712401] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bf7979-43bc-4f72-81bc-650e8b24e880 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.748010] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9fd9fc-a184-4eec-8c3a-2f19e02e9e5f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.755531] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272192, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.49005} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.757647] env[62235]: INFO nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098/OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098.vmdk to [datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32/ede39f76-6301-4e86-86d4-85e79bd11b32.vmdk. [ 1078.757942] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Cleaning up location [datastore2] OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1078.758063] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e1c71327-c536-404c-b260-fd6c6769e098 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.758354] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d42ade9e-5fb5-47a0-ab6a-ba8a93b35524 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.760868] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7a67d8-951c-4785-94b9-b36f68270cdc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.775088] env[62235]: DEBUG nova.compute.provider_tree [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.777421] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1078.777421] env[62235]: value = "task-1272194" [ 1078.777421] env[62235]: _type = "Task" [ 1078.777421] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.785378] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.035694] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 2068403d-5870-4daa-b4d6-acaf39209488] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.279121] env[62235]: DEBUG nova.scheduler.client.report [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.291789] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.050268} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.292098] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.292311] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32/ede39f76-6301-4e86-86d4-85e79bd11b32.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.292593] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32/ede39f76-6301-4e86-86d4-85e79bd11b32.vmdk to [datastore2] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1079.292880] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88f4695e-7c0c-450c-a141-1c4c960b9ab2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.299902] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1079.299902] env[62235]: value = "task-1272195" [ 1079.299902] env[62235]: _type = "Task" [ 1079.299902] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.307216] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.432678] env[62235]: DEBUG nova.network.neutron [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Port c81e1539-5f47-4bca-bd85-f3ed740feeeb binding to destination host cpu-1 is already ACTIVE {{(pid=62235) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1079.432947] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.433149] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.433339] env[62235]: DEBUG nova.network.neutron [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1079.538887] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 062bbe32-3edd-4b06-b7d9-882e798b3a7f] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.787943] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.191s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.788521] env[62235]: DEBUG nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.810963] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.042543] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ecb104db-2172-4b09-867a-7a2813e03fa5] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.177853] env[62235]: DEBUG nova.network.neutron [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.293566] env[62235]: DEBUG nova.compute.utils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1080.294995] env[62235]: DEBUG nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1080.295179] env[62235]: DEBUG nova.network.neutron [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1080.310611] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.332061] env[62235]: DEBUG nova.policy [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93109c6c348a4975b8c144c37e8c4f82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab1aacbb5a8746769c04f2df5f99d01a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1080.545791] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3ba0a88f-6de9-46ed-9c98-306e13dcc3f4] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.680423] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.798110] env[62235]: DEBUG nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1080.810666] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.854950] env[62235]: DEBUG nova.network.neutron [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Successfully created port: 168418b4-2660-419a-883f-f16e2c1501ca {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1081.049844] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b5e4c6b2-4583-4a20-b570-99699063e172] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.183469] env[62235]: DEBUG nova.compute.manager [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62235) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1081.183724] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.183972] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.314515] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.554263] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 44ed8739-ad20-41a3-abe1-38176c1bee25] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.689090] env[62235]: DEBUG nova.objects.instance [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'migration_context' on Instance uuid ef220f90-5581-4a51-b308-12da850782a7 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.813741] env[62235]: DEBUG nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1081.825605] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.840710] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.840997] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.841206] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.841412] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.841568] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.841725] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.841942] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.842141] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.842339] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.842513] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.842693] env[62235]: DEBUG nova.virt.hardware [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.843646] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b00e623-e943-4c34-b3c1-f8cd582fd1b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.852076] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ab2388-f09c-4b3c-9b7a-7efded4cf646 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.061360] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 507fa559-1889-4f6e-81ee-93eb14d134b2] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.317559] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.396427] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7853eec6-d3f5-418c-b08c-cafeab94bf2a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.404576] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed9aae9-65c8-411c-bd97-89a93565780a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.435317] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b9105a-5cee-4fad-8422-32bd1f9501a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.444144] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f031855-1627-40e9-b2e8-748c1ebd6583 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.448980] env[62235]: DEBUG nova.compute.manager [req-0baf896b-ba1d-48dc-b243-d27197f50ef1 req-fda52a20-d031-42ea-aeeb-c74caacd5f65 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Received event network-vif-plugged-168418b4-2660-419a-883f-f16e2c1501ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.449223] env[62235]: DEBUG oslo_concurrency.lockutils [req-0baf896b-ba1d-48dc-b243-d27197f50ef1 req-fda52a20-d031-42ea-aeeb-c74caacd5f65 service nova] Acquiring lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.449436] env[62235]: DEBUG oslo_concurrency.lockutils [req-0baf896b-ba1d-48dc-b243-d27197f50ef1 req-fda52a20-d031-42ea-aeeb-c74caacd5f65 service nova] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.449608] env[62235]: DEBUG oslo_concurrency.lockutils [req-0baf896b-ba1d-48dc-b243-d27197f50ef1 req-fda52a20-d031-42ea-aeeb-c74caacd5f65 service nova] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.449779] env[62235]: DEBUG nova.compute.manager [req-0baf896b-ba1d-48dc-b243-d27197f50ef1 req-fda52a20-d031-42ea-aeeb-c74caacd5f65 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] No waiting events found dispatching network-vif-plugged-168418b4-2660-419a-883f-f16e2c1501ca {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1082.449947] env[62235]: WARNING nova.compute.manager [req-0baf896b-ba1d-48dc-b243-d27197f50ef1 req-fda52a20-d031-42ea-aeeb-c74caacd5f65 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Received unexpected event network-vif-plugged-168418b4-2660-419a-883f-f16e2c1501ca for instance with vm_state building and task_state spawning. [ 1082.460367] env[62235]: DEBUG nova.compute.provider_tree [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.511754] env[62235]: DEBUG nova.network.neutron [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Successfully updated port: 168418b4-2660-419a-883f-f16e2c1501ca {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.564289] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: c571a362-9b55-42d3-a019-1695c1caf248] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.817139] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272195, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.351822} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.817416] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ede39f76-6301-4e86-86d4-85e79bd11b32/ede39f76-6301-4e86-86d4-85e79bd11b32.vmdk to [datastore2] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1082.818249] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8570702f-13ba-4598-a3dc-5a65bfd86123 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.839415] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.839675] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a73cb4ad-f2c5-4cc2-9dd1-32d23e1725ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.860278] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1082.860278] env[62235]: value = "task-1272196" [ 1082.860278] env[62235]: _type = "Task" [ 1082.860278] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.871854] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272196, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.964255] env[62235]: DEBUG nova.scheduler.client.report [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.014498] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "refresh_cache-f1ef38cf-4f4b-4c53-a69d-5c03493a411f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.014674] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "refresh_cache-f1ef38cf-4f4b-4c53-a69d-5c03493a411f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.015047] env[62235]: DEBUG nova.network.neutron [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1083.067738] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 15d41c0f-045d-4e1c-88b5-1841f99f2e74] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.370608] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272196, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.545297] env[62235]: DEBUG nova.network.neutron [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1083.572708] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b7801744-5552-49b8-a95d-41b67bdf2380] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.716717] env[62235]: DEBUG nova.network.neutron [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Updating instance_info_cache with network_info: [{"id": "168418b4-2660-419a-883f-f16e2c1501ca", "address": "fa:16:3e:2e:2e:03", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap168418b4-26", "ovs_interfaceid": "168418b4-2660-419a-883f-f16e2c1501ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.872913] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272196, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.977153] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.793s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.076388] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 9474de53-1964-4a52-8783-2bac5773b6c9] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.219271] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "refresh_cache-f1ef38cf-4f4b-4c53-a69d-5c03493a411f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.219613] env[62235]: DEBUG nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Instance network_info: |[{"id": "168418b4-2660-419a-883f-f16e2c1501ca", "address": "fa:16:3e:2e:2e:03", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap168418b4-26", "ovs_interfaceid": "168418b4-2660-419a-883f-f16e2c1501ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1084.220074] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:2e:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73eeba7c-29e1-4fdf-82b3-d62e63e86051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '168418b4-2660-419a-883f-f16e2c1501ca', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.227545] env[62235]: DEBUG oslo.service.loopingcall [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.227762] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1084.228302] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8c915a9-7a51-42db-9287-72428e5dcdab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.250718] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.250718] env[62235]: value = "task-1272197" [ 1084.250718] env[62235]: _type = "Task" [ 1084.250718] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.259185] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272197, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.372542] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272196, 'name': ReconfigVM_Task, 'duration_secs': 1.061352} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.372845] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 738750ad-931e-45e4-b0c6-8232fe1b9f46/738750ad-931e-45e4-b0c6-8232fe1b9f46.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.373553] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88034379-451c-4854-aa28-93b95a0d88d5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.380892] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1084.380892] env[62235]: value = "task-1272198" [ 1084.380892] env[62235]: _type = "Task" [ 1084.380892] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.389192] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272198, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.466590] env[62235]: DEBUG nova.compute.manager [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Received event network-changed-168418b4-2660-419a-883f-f16e2c1501ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.466801] env[62235]: DEBUG nova.compute.manager [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Refreshing instance network info cache due to event network-changed-168418b4-2660-419a-883f-f16e2c1501ca. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1084.467029] env[62235]: DEBUG oslo_concurrency.lockutils [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] Acquiring lock "refresh_cache-f1ef38cf-4f4b-4c53-a69d-5c03493a411f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.467187] env[62235]: DEBUG oslo_concurrency.lockutils [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] Acquired lock "refresh_cache-f1ef38cf-4f4b-4c53-a69d-5c03493a411f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.467351] env[62235]: DEBUG nova.network.neutron [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Refreshing network info cache for port 168418b4-2660-419a-883f-f16e2c1501ca {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1084.580318] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b126a2c9-62e9-49a1-91ae-7a703e5fa490] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.761704] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272197, 'name': CreateVM_Task, 'duration_secs': 0.353356} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.761881] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1084.762625] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.762799] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.763193] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1084.763451] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc991e27-d4cd-4d56-b4bc-74fc6c20ca31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.768306] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1084.768306] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f367a1-0915-9dfc-29f7-ae13930179ca" [ 1084.768306] env[62235]: _type = "Task" [ 1084.768306] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.776660] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f367a1-0915-9dfc-29f7-ae13930179ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.890737] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272198, 'name': Rename_Task, 'duration_secs': 0.18837} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.891048] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.891311] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d9cddb1-ca51-45db-a8f6-33a83a6dd1b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.898152] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1084.898152] env[62235]: value = "task-1272199" [ 1084.898152] env[62235]: _type = "Task" [ 1084.898152] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.905538] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272199, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.083742] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 9f5fcb83-bb6f-4815-b842-a37b2dbf26e9] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.163870] env[62235]: DEBUG nova.network.neutron [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Updated VIF entry in instance network info cache for port 168418b4-2660-419a-883f-f16e2c1501ca. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1085.164339] env[62235]: DEBUG nova.network.neutron [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Updating instance_info_cache with network_info: [{"id": "168418b4-2660-419a-883f-f16e2c1501ca", "address": "fa:16:3e:2e:2e:03", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap168418b4-26", "ovs_interfaceid": "168418b4-2660-419a-883f-f16e2c1501ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.280585] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f367a1-0915-9dfc-29f7-ae13930179ca, 'name': SearchDatastore_Task, 'duration_secs': 0.010366} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.280912] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.281181] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.281432] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.281586] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.281774] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.282058] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c0d2d477-112c-4548-8c28-49d4fb1add34 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.291048] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.291284] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1085.292078] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e7023f0-0650-4c11-85c8-ecb02f3f58cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.297601] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1085.297601] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a66f87-1bf4-fdd7-de5d-e4b7cf5e543f" [ 1085.297601] env[62235]: _type = "Task" [ 1085.297601] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.306620] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a66f87-1bf4-fdd7-de5d-e4b7cf5e543f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.409101] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272199, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.518326] env[62235]: INFO nova.compute.manager [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Swapping old allocation on dict_keys(['108a4390-ff68-4048-b61d-c7a9614ddc4f']) held by migration 184e870e-abd3-4f70-b21d-002e64b1d06c for instance [ 1085.539705] env[62235]: DEBUG nova.scheduler.client.report [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Overwriting current allocation {'allocations': {'108a4390-ff68-4048-b61d-c7a9614ddc4f': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 132}}, 'project_id': 'c943a445f67f4021bad11723e0836537', 'user_id': '738ee55c831c48f894efbcef71bc20a4', 'consumer_generation': 1} on consumer ef220f90-5581-4a51-b308-12da850782a7 {{(pid=62235) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1085.587814] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 00825c9e-b398-4edc-a5ed-e4b5b476af1c] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.616109] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.616329] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.616526] env[62235]: DEBUG nova.network.neutron [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1085.667316] env[62235]: DEBUG oslo_concurrency.lockutils [req-393df750-d5d1-402f-8ba4-04146f90963a req-da53b6e5-61a4-4639-9477-e269cef77fb0 service nova] Releasing lock "refresh_cache-f1ef38cf-4f4b-4c53-a69d-5c03493a411f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.808574] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a66f87-1bf4-fdd7-de5d-e4b7cf5e543f, 'name': SearchDatastore_Task, 'duration_secs': 0.011566} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.809386] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f28766d8-f735-489c-ab43-71b354f5c9e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.815570] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1085.815570] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d63ec1-5871-52e0-512d-f4d20d23e872" [ 1085.815570] env[62235]: _type = "Task" [ 1085.815570] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.824034] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d63ec1-5871-52e0-512d-f4d20d23e872, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.908136] env[62235]: DEBUG oslo_vmware.api [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272199, 'name': PowerOnVM_Task, 'duration_secs': 0.552572} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.908420] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.017661] env[62235]: DEBUG nova.compute.manager [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.018607] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cef6337-def6-4695-a745-9125b97f0196 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.091199] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 80a803d1-c034-42b6-a15d-9fc3893953ad] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.326352] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d63ec1-5871-52e0-512d-f4d20d23e872, 'name': SearchDatastore_Task, 'duration_secs': 0.009934} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.326650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.326909] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] f1ef38cf-4f4b-4c53-a69d-5c03493a411f/f1ef38cf-4f4b-4c53-a69d-5c03493a411f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1086.327190] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1769a7e8-1bba-41ae-a2bd-85b8ef9a5552 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.335753] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1086.335753] env[62235]: value = "task-1272200" [ 1086.335753] env[62235]: _type = "Task" [ 1086.335753] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.343962] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272200, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.345843] env[62235]: DEBUG nova.network.neutron [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [{"id": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "address": "fa:16:3e:1d:13:1c", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc81e1539-5f", "ovs_interfaceid": "c81e1539-5f47-4bca-bd85-f3ed740feeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.537800] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c4455a56-f0ba-4138-ac8f-69bd462a8d0f tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.342s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.594804] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 06ec6b9c-bef0-4f96-a185-4315961de7f7] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.846558] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272200, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.414521} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.846822] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] f1ef38cf-4f4b-4c53-a69d-5c03493a411f/f1ef38cf-4f4b-4c53-a69d-5c03493a411f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1086.847055] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1086.847315] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c67e0b0-93aa-4eb8-a6bc-e35d5133d744 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.849324] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-ef220f90-5581-4a51-b308-12da850782a7" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.849936] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.850231] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ea32ef2-8c32-430c-bc76-2efd5ed09e5e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.857829] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1086.857829] env[62235]: value = "task-1272202" [ 1086.857829] env[62235]: _type = "Task" [ 1086.857829] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.859206] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1086.859206] env[62235]: value = "task-1272201" [ 1086.859206] env[62235]: _type = "Task" [ 1086.859206] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.871241] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.874259] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272202, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.098299] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 876b7602-ccf0-4aed-90e2-d07fbf548fb6] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.374602] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272201, 'name': PowerOffVM_Task, 'duration_secs': 0.245209} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.374826] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272202, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066605} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.375042] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.375705] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1087.375924] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1087.376102] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.376297] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1087.376448] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.376599] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1087.376818] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1087.376974] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1087.377164] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1087.377335] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1087.377514] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1087.382322] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1087.382586] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a13d5a2-7e0a-48e8-8d9b-c7ea91ce002d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.392984] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f729c7fe-1c7e-49f3-8080-cc26b50ec0ca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.417332] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] f1ef38cf-4f4b-4c53-a69d-5c03493a411f/f1ef38cf-4f4b-4c53-a69d-5c03493a411f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.418826] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aeb3b17b-2da5-44d1-b368-89a96ad0a1b6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.433925] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1087.433925] env[62235]: value = "task-1272203" [ 1087.433925] env[62235]: _type = "Task" [ 1087.433925] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.439744] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1087.439744] env[62235]: value = "task-1272204" [ 1087.439744] env[62235]: _type = "Task" [ 1087.439744] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.443307] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272203, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.450886] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272204, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.601738] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b5a1da09-2137-4434-a1a4-174b196b61b0] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.946941] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272203, 'name': ReconfigVM_Task, 'duration_secs': 0.460356} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.950621] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212f9917-29a4-4753-96ef-fd6ef9127f34 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.960378] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272204, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.976464] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1087.976806] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1087.976977] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.977189] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1087.977341] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.977493] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1087.977703] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1087.977869] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1087.978098] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1087.978282] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1087.978462] env[62235]: DEBUG nova.virt.hardware [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1087.979326] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-518606d5-f57c-4db4-bbb2-e9b959b2762d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.986137] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1087.986137] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c38d2d-a790-8748-a9c4-6f401d48d16a" [ 1087.986137] env[62235]: _type = "Task" [ 1087.986137] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.995745] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c38d2d-a790-8748-a9c4-6f401d48d16a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.106038] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ced2aa67-ce75-4af5-9c55-773835d05733] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.460149] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272204, 'name': ReconfigVM_Task, 'duration_secs': 0.896769} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.460207] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Reconfigured VM instance instance-00000067 to attach disk [datastore1] f1ef38cf-4f4b-4c53-a69d-5c03493a411f/f1ef38cf-4f4b-4c53-a69d-5c03493a411f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.460977] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfac22b3-7794-4d63-bb10-70eecd3dbfbe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.469245] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1088.469245] env[62235]: value = "task-1272205" [ 1088.469245] env[62235]: _type = "Task" [ 1088.469245] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.477834] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272205, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.495899] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c38d2d-a790-8748-a9c4-6f401d48d16a, 'name': SearchDatastore_Task, 'duration_secs': 0.00824} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.502263] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1088.502622] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a66ad9e-a41a-4a76-a1ef-305914ae1b90 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.521828] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1088.521828] env[62235]: value = "task-1272206" [ 1088.521828] env[62235]: _type = "Task" [ 1088.521828] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.530186] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272206, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.609606] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d165f549-1759-41bf-86fd-b77793bd49c0] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.979628] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272205, 'name': Rename_Task, 'duration_secs': 0.16526} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.979852] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.980220] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5aaffbb7-7cd8-4dbf-954d-436aee0f1585 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.987145] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1088.987145] env[62235]: value = "task-1272207" [ 1088.987145] env[62235]: _type = "Task" [ 1088.987145] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.994972] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.031774] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272206, 'name': ReconfigVM_Task, 'duration_secs': 0.164285} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.032405] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1089.032877] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f49761-490a-4958-9d4b-f9801260cf14 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.055021] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.055441] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dec52d13-6880-4a1e-8a7f-06861aac8845 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.075410] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1089.075410] env[62235]: value = "task-1272208" [ 1089.075410] env[62235]: _type = "Task" [ 1089.075410] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.085792] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272208, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.112659] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0e9182f8-81e1-4ca2-97b6-0066faba0b22] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.271991] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "e6421e2f-17cf-4312-a122-067cd032c066" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.272271] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.497950] env[62235]: DEBUG oslo_vmware.api [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272207, 'name': PowerOnVM_Task, 'duration_secs': 0.465385} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.498293] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1089.498579] env[62235]: INFO nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Took 7.68 seconds to spawn the instance on the hypervisor. [ 1089.498835] env[62235]: DEBUG nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.499642] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce499883-6d52-4ab4-ba34-8bba09862e56 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.585953] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272208, 'name': ReconfigVM_Task, 'duration_secs': 0.298464} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.586271] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Reconfigured VM instance instance-00000061 to attach disk [datastore2] ef220f90-5581-4a51-b308-12da850782a7/ef220f90-5581-4a51-b308-12da850782a7.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.587125] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75c2e85-2044-403a-bee8-e12b53ec0128 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.607055] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91015a5-cdf4-4be4-9632-af6a16aff51c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.624322] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 590440de-4b3b-4019-b0c3-58b5ad329c88] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.626620] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88299a3f-ff28-4130-aa62-e041b8c845e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.644691] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e48096-07d2-453d-9d82-aa2bd11848b8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.652326] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.652563] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b23ffdd6-5a77-4ac5-a844-f7ac4bd50959 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.659350] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1089.659350] env[62235]: value = "task-1272209" [ 1089.659350] env[62235]: _type = "Task" [ 1089.659350] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.666392] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272209, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.774428] env[62235]: DEBUG nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1090.018048] env[62235]: INFO nova.compute.manager [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Took 12.44 seconds to build instance. [ 1090.130117] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 63e16fd4-3a5a-4231-9013-c6a01f0eab80] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.170273] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272209, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.299150] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.299631] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.302543] env[62235]: INFO nova.compute.claims [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1090.519955] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09dcba7b-f415-48db-a729-250b828421a8 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.950s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.633663] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 3ea7db36-ebde-4523-b803-ad121a00269b] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.670248] env[62235]: DEBUG oslo_vmware.api [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272209, 'name': PowerOnVM_Task, 'duration_secs': 0.705756} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.670533] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1090.925274] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.925507] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.925725] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.925914] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.926109] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.928198] env[62235]: INFO nova.compute.manager [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Terminating instance [ 1090.930089] env[62235]: DEBUG nova.compute.manager [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.930295] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1090.931523] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db280f19-3ca7-4ee4-ba82-f0d175c1611f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.939801] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.940045] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08154644-75ff-4ac7-bdcf-7be21aebdebc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.946453] env[62235]: DEBUG oslo_vmware.api [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1090.946453] env[62235]: value = "task-1272210" [ 1090.946453] env[62235]: _type = "Task" [ 1090.946453] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.954914] env[62235]: DEBUG oslo_vmware.api [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.137420] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b4a3777f-2ee5-464b-84eb-0722d59f2b44] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.398596] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01d4cbf-8922-4c4c-8e53-b7ceda11193e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.408220] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5be43d-7eea-4e54-bc96-1828103c3a55 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.438136] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2084a6bc-0dd3-4a24-96ea-311099ad3940 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.445871] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccdbf2f-298f-4b0a-89d5-ae8c6181272d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.461806] env[62235]: DEBUG nova.compute.provider_tree [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.465833] env[62235]: DEBUG oslo_vmware.api [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272210, 'name': PowerOffVM_Task, 'duration_secs': 0.264422} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.466123] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.466311] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.466760] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16d7cf06-0d0c-4869-bb62-5e0a3e7d4682 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.531489] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.531737] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.531913] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleting the datastore file [datastore1] f1ef38cf-4f4b-4c53-a69d-5c03493a411f {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.532214] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f42f0693-55cc-4eee-90a5-f97038319460 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.541895] env[62235]: DEBUG oslo_vmware.api [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1091.541895] env[62235]: value = "task-1272212" [ 1091.541895] env[62235]: _type = "Task" [ 1091.541895] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.550155] env[62235]: DEBUG oslo_vmware.api [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272212, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.640481] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ec1830c7-7095-45d1-a188-c2993c60035e] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.681846] env[62235]: INFO nova.compute.manager [None req-e81448e7-eed2-43c8-8a15-e77b6714d7fe tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance to original state: 'active' [ 1091.967348] env[62235]: DEBUG nova.scheduler.client.report [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.053949] env[62235]: DEBUG oslo_vmware.api [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272212, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126244} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.054246] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.054443] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.054616] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.054791] env[62235]: INFO nova.compute.manager [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1092.055054] env[62235]: DEBUG oslo.service.loopingcall [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.055283] env[62235]: DEBUG nova.compute.manager [-] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1092.055382] env[62235]: DEBUG nova.network.neutron [-] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1092.143851] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: afa35ae0-40c6-4eaa-80f5-c1ada27f1871] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.313936] env[62235]: DEBUG nova.compute.manager [req-f856ec9a-feda-416b-8c00-00798b383801 req-0ef97c3b-88b7-45e0-bc3e-4cac0224e649 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Received event network-vif-deleted-168418b4-2660-419a-883f-f16e2c1501ca {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.315044] env[62235]: INFO nova.compute.manager [req-f856ec9a-feda-416b-8c00-00798b383801 req-0ef97c3b-88b7-45e0-bc3e-4cac0224e649 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Neutron deleted interface 168418b4-2660-419a-883f-f16e2c1501ca; detaching it from the instance and deleting it from the info cache [ 1092.315044] env[62235]: DEBUG nova.network.neutron [req-f856ec9a-feda-416b-8c00-00798b383801 req-0ef97c3b-88b7-45e0-bc3e-4cac0224e649 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.472566] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.473217] env[62235]: DEBUG nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1092.646800] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: d6324b5d-ec70-4bb4-b3c6-b1524abd0034] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.791043] env[62235]: DEBUG nova.network.neutron [-] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.817539] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1446a13b-a729-4f58-a968-27afa1d71e2b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.828409] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced5f7fc-1136-440a-9d37-11a93f92b947 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.856932] env[62235]: DEBUG nova.compute.manager [req-f856ec9a-feda-416b-8c00-00798b383801 req-0ef97c3b-88b7-45e0-bc3e-4cac0224e649 service nova] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Detach interface failed, port_id=168418b4-2660-419a-883f-f16e2c1501ca, reason: Instance f1ef38cf-4f4b-4c53-a69d-5c03493a411f could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1092.978198] env[62235]: DEBUG nova.compute.utils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1092.979450] env[62235]: DEBUG nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1092.979627] env[62235]: DEBUG nova.network.neutron [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1093.018712] env[62235]: DEBUG nova.policy [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91b451aa2fe74f6eabee71b0f4f90f0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cd98f2307d24723b3aeb481b535ace9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1093.150380] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: dd7268d1-1da5-4e5f-9d45-816df147b564] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.263661] env[62235]: DEBUG nova.network.neutron [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Successfully created port: 5357f0fa-5c91-459c-8fc1-4117e140bedb {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1093.293415] env[62235]: INFO nova.compute.manager [-] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Took 1.24 seconds to deallocate network for instance. [ 1093.312959] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "ef220f90-5581-4a51-b308-12da850782a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.313256] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.313470] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "ef220f90-5581-4a51-b308-12da850782a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.313659] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.313932] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.315782] env[62235]: INFO nova.compute.manager [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Terminating instance [ 1093.317544] env[62235]: DEBUG nova.compute.manager [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1093.317744] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.318623] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15502caf-d7e8-4e41-a0e3-271b9c79aa01 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.327978] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1093.328257] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f48f8e2-cdb0-4e66-a368-ed4f746aa400 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.336842] env[62235]: DEBUG oslo_vmware.api [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1093.336842] env[62235]: value = "task-1272213" [ 1093.336842] env[62235]: _type = "Task" [ 1093.336842] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.346067] env[62235]: DEBUG oslo_vmware.api [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272213, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.482792] env[62235]: DEBUG nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1093.654067] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8559690a-bdd3-4ec3-98e1-dfc78c6db8ba] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.799824] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.799989] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.800236] env[62235]: DEBUG nova.objects.instance [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lazy-loading 'resources' on Instance uuid f1ef38cf-4f4b-4c53-a69d-5c03493a411f {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.846782] env[62235]: DEBUG oslo_vmware.api [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272213, 'name': PowerOffVM_Task, 'duration_secs': 0.220538} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.847081] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.847261] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.847520] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d680bcd-0c19-4a3c-87ae-ac593d3c98a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.110266] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1094.110499] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1094.110688] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleting the datastore file [datastore2] ef220f90-5581-4a51-b308-12da850782a7 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.111034] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98179bb7-c239-4e01-ba0d-e0afb583dfa9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.118311] env[62235]: DEBUG oslo_vmware.api [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1094.118311] env[62235]: value = "task-1272215" [ 1094.118311] env[62235]: _type = "Task" [ 1094.118311] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.127330] env[62235]: DEBUG oslo_vmware.api [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272215, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.157435] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.157735] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Cleaning up deleted instances with incomplete migration {{(pid=62235) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1094.395532] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce208d1-3647-4232-9ccf-247655e66601 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.403201] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077a4cbf-3b11-4772-8898-39574e3b9721 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.434554] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882098a9-b9ab-4696-a948-c57b5ce3c950 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.441992] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494a030f-4527-4d8e-bfb6-e6ab6ba656d6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.455355] env[62235]: DEBUG nova.compute.provider_tree [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.493935] env[62235]: DEBUG nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1094.519076] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.519353] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.519517] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.519705] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.519861] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.520032] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.520258] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.520424] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.520596] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.520838] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.520986] env[62235]: DEBUG nova.virt.hardware [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.521856] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1f1ead-44e1-4fc9-be46-49034879917a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.531657] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee92726a-de19-4e94-837b-83f6b1514434 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.629091] env[62235]: DEBUG oslo_vmware.api [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272215, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147767} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.629384] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.629606] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1094.629813] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1094.630034] env[62235]: INFO nova.compute.manager [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: ef220f90-5581-4a51-b308-12da850782a7] Took 1.31 seconds to destroy the instance on the hypervisor. [ 1094.630306] env[62235]: DEBUG oslo.service.loopingcall [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.630500] env[62235]: DEBUG nova.compute.manager [-] [instance: ef220f90-5581-4a51-b308-12da850782a7] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1094.630593] env[62235]: DEBUG nova.network.neutron [-] [instance: ef220f90-5581-4a51-b308-12da850782a7] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1094.642184] env[62235]: DEBUG nova.compute.manager [req-ccd7f2fc-c545-46ad-a614-7a7fd32402a6 req-3621d50d-17e5-42d2-a1cf-f041fe433ca8 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Received event network-vif-plugged-5357f0fa-5c91-459c-8fc1-4117e140bedb {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.642412] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccd7f2fc-c545-46ad-a614-7a7fd32402a6 req-3621d50d-17e5-42d2-a1cf-f041fe433ca8 service nova] Acquiring lock "e6421e2f-17cf-4312-a122-067cd032c066-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.642944] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccd7f2fc-c545-46ad-a614-7a7fd32402a6 req-3621d50d-17e5-42d2-a1cf-f041fe433ca8 service nova] Lock "e6421e2f-17cf-4312-a122-067cd032c066-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.643382] env[62235]: DEBUG oslo_concurrency.lockutils [req-ccd7f2fc-c545-46ad-a614-7a7fd32402a6 req-3621d50d-17e5-42d2-a1cf-f041fe433ca8 service nova] Lock "e6421e2f-17cf-4312-a122-067cd032c066-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.643619] env[62235]: DEBUG nova.compute.manager [req-ccd7f2fc-c545-46ad-a614-7a7fd32402a6 req-3621d50d-17e5-42d2-a1cf-f041fe433ca8 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] No waiting events found dispatching network-vif-plugged-5357f0fa-5c91-459c-8fc1-4117e140bedb {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1094.643829] env[62235]: WARNING nova.compute.manager [req-ccd7f2fc-c545-46ad-a614-7a7fd32402a6 req-3621d50d-17e5-42d2-a1cf-f041fe433ca8 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Received unexpected event network-vif-plugged-5357f0fa-5c91-459c-8fc1-4117e140bedb for instance with vm_state building and task_state spawning. [ 1094.661780] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.736306] env[62235]: DEBUG nova.network.neutron [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Successfully updated port: 5357f0fa-5c91-459c-8fc1-4117e140bedb {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1094.957997] env[62235]: DEBUG nova.scheduler.client.report [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.239783] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-e6421e2f-17cf-4312-a122-067cd032c066" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.239783] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-e6421e2f-17cf-4312-a122-067cd032c066" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.239783] env[62235]: DEBUG nova.network.neutron [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1095.331958] env[62235]: DEBUG nova.compute.manager [req-110551db-8d73-4b04-8455-848e16840f0c req-bd70dc07-cd17-4702-89a0-8a1dc148f84f service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Received event network-vif-deleted-c81e1539-5f47-4bca-bd85-f3ed740feeeb {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.332204] env[62235]: INFO nova.compute.manager [req-110551db-8d73-4b04-8455-848e16840f0c req-bd70dc07-cd17-4702-89a0-8a1dc148f84f service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Neutron deleted interface c81e1539-5f47-4bca-bd85-f3ed740feeeb; detaching it from the instance and deleting it from the info cache [ 1095.332356] env[62235]: DEBUG nova.network.neutron [req-110551db-8d73-4b04-8455-848e16840f0c req-bd70dc07-cd17-4702-89a0-8a1dc148f84f service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.462703] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.663s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.487342] env[62235]: INFO nova.scheduler.client.report [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleted allocations for instance f1ef38cf-4f4b-4c53-a69d-5c03493a411f [ 1095.771215] env[62235]: DEBUG nova.network.neutron [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1095.807475] env[62235]: DEBUG nova.network.neutron [-] [instance: ef220f90-5581-4a51-b308-12da850782a7] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.836434] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3b6b221-e725-4a01-87a0-011f370a91a6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.849232] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b24aaf-2c7f-4453-8f8a-8ade1a3eab9a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.881736] env[62235]: DEBUG nova.compute.manager [req-110551db-8d73-4b04-8455-848e16840f0c req-bd70dc07-cd17-4702-89a0-8a1dc148f84f service nova] [instance: ef220f90-5581-4a51-b308-12da850782a7] Detach interface failed, port_id=c81e1539-5f47-4bca-bd85-f3ed740feeeb, reason: Instance ef220f90-5581-4a51-b308-12da850782a7 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1095.947625] env[62235]: DEBUG nova.network.neutron [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Updating instance_info_cache with network_info: [{"id": "5357f0fa-5c91-459c-8fc1-4117e140bedb", "address": "fa:16:3e:f7:39:e9", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5357f0fa-5c", "ovs_interfaceid": "5357f0fa-5c91-459c-8fc1-4117e140bedb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.995976] env[62235]: DEBUG oslo_concurrency.lockutils [None req-f0e7b7f8-089b-4219-907d-1fb7636cdbde tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "f1ef38cf-4f4b-4c53-a69d-5c03493a411f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.070s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.311936] env[62235]: INFO nova.compute.manager [-] [instance: ef220f90-5581-4a51-b308-12da850782a7] Took 1.68 seconds to deallocate network for instance. [ 1096.450141] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-e6421e2f-17cf-4312-a122-067cd032c066" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.450501] env[62235]: DEBUG nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Instance network_info: |[{"id": "5357f0fa-5c91-459c-8fc1-4117e140bedb", "address": "fa:16:3e:f7:39:e9", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5357f0fa-5c", "ovs_interfaceid": "5357f0fa-5c91-459c-8fc1-4117e140bedb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1096.450961] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:39:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5357f0fa-5c91-459c-8fc1-4117e140bedb', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1096.458501] env[62235]: DEBUG oslo.service.loopingcall [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1096.458724] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1096.458947] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-608e6f9d-7b76-44d0-bb75-6bbc23fe4617 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.479059] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1096.479059] env[62235]: value = "task-1272216" [ 1096.479059] env[62235]: _type = "Task" [ 1096.479059] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.486860] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272216, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.669768] env[62235]: DEBUG nova.compute.manager [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Received event network-changed-5357f0fa-5c91-459c-8fc1-4117e140bedb {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1096.669985] env[62235]: DEBUG nova.compute.manager [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Refreshing instance network info cache due to event network-changed-5357f0fa-5c91-459c-8fc1-4117e140bedb. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1096.670231] env[62235]: DEBUG oslo_concurrency.lockutils [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] Acquiring lock "refresh_cache-e6421e2f-17cf-4312-a122-067cd032c066" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.670408] env[62235]: DEBUG oslo_concurrency.lockutils [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] Acquired lock "refresh_cache-e6421e2f-17cf-4312-a122-067cd032c066" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.670612] env[62235]: DEBUG nova.network.neutron [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Refreshing network info cache for port 5357f0fa-5c91-459c-8fc1-4117e140bedb {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1096.818217] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.818494] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.818695] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.839076] env[62235]: INFO nova.scheduler.client.report [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted allocations for instance ef220f90-5581-4a51-b308-12da850782a7 [ 1096.989185] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272216, 'name': CreateVM_Task, 'duration_secs': 0.48773} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.989332] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1096.990078] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.990276] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.990622] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1096.990920] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcdc2bd8-4b6c-4ac5-b3a4-72b8e8543b0e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.996295] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1096.996295] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523286cb-7fec-4c4f-a1d2-b38daca18d08" [ 1096.996295] env[62235]: _type = "Task" [ 1096.996295] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.004982] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523286cb-7fec-4c4f-a1d2-b38daca18d08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.346863] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9717dbae-c655-481d-8ea4-aa6409a7f1b2 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "ef220f90-5581-4a51-b308-12da850782a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.033s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.449029] env[62235]: DEBUG nova.network.neutron [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Updated VIF entry in instance network info cache for port 5357f0fa-5c91-459c-8fc1-4117e140bedb. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1097.449456] env[62235]: DEBUG nova.network.neutron [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Updating instance_info_cache with network_info: [{"id": "5357f0fa-5c91-459c-8fc1-4117e140bedb", "address": "fa:16:3e:f7:39:e9", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5357f0fa-5c", "ovs_interfaceid": "5357f0fa-5c91-459c-8fc1-4117e140bedb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.508746] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523286cb-7fec-4c4f-a1d2-b38daca18d08, 'name': SearchDatastore_Task, 'duration_secs': 0.010099} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.509089] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.509332] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.509563] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.509707] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.509885] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.510166] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6638ea54-183e-40d7-9004-f247c89178dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.519952] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.520687] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1097.521516] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecac41fa-5551-45d0-8913-3d038150d5d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.527493] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1097.527493] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5296c3dd-6cb1-bcbc-b7dd-28b6c0800fe5" [ 1097.527493] env[62235]: _type = "Task" [ 1097.527493] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.537104] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5296c3dd-6cb1-bcbc-b7dd-28b6c0800fe5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.953685] env[62235]: DEBUG oslo_concurrency.lockutils [req-626f465b-a93e-4669-b92a-ffe04c141ce6 req-17f41ef8-92d5-4da2-a80b-11cc7ef73f96 service nova] Releasing lock "refresh_cache-e6421e2f-17cf-4312-a122-067cd032c066" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.038539] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5296c3dd-6cb1-bcbc-b7dd-28b6c0800fe5, 'name': SearchDatastore_Task, 'duration_secs': 0.008921} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.039384] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e82e1295-5ea9-4af8-b31d-6ff4fd2d9060 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.044903] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1098.044903] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52569337-55bf-699b-eba4-7f2563822149" [ 1098.044903] env[62235]: _type = "Task" [ 1098.044903] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.053157] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52569337-55bf-699b-eba4-7f2563822149, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.263255] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "f7ef033e-ab43-4b89-b331-cb234a373991" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.263620] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.555777] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52569337-55bf-699b-eba4-7f2563822149, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.556102] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.556314] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1098.556575] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35870149-30f5-4ea7-a9fd-772d95f8c24e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.564309] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1098.564309] env[62235]: value = "task-1272217" [ 1098.564309] env[62235]: _type = "Task" [ 1098.564309] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.572319] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.768056] env[62235]: DEBUG nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1099.076216] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.419243} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.076495] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1099.076713] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1099.077076] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c938b30-050e-4cf1-8a6c-484cfe163099 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.084836] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1099.084836] env[62235]: value = "task-1272218" [ 1099.084836] env[62235]: _type = "Task" [ 1099.084836] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.094123] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272218, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.289519] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.289607] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.291182] env[62235]: INFO nova.compute.claims [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1099.596164] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272218, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090914} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.596164] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.597131] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4e7e61-51c3-448e-aec7-b7d5a1e89f6a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.619574] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.619896] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1b0c39f-d3b8-4c2b-8843-30df47541197 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.640646] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1099.640646] env[62235]: value = "task-1272219" [ 1099.640646] env[62235]: _type = "Task" [ 1099.640646] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.648623] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272219, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.152089] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272219, 'name': ReconfigVM_Task, 'duration_secs': 0.288425} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.152089] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Reconfigured VM instance instance-00000068 to attach disk [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.152463] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a13a0464-c54c-4fb4-bdfe-00436fb9e908 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.159320] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1100.159320] env[62235]: value = "task-1272220" [ 1100.159320] env[62235]: _type = "Task" [ 1100.159320] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.167112] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272220, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.378415] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efee69c7-84ca-4a92-bdee-ffa5156f4813 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.386538] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f9635f-8cb5-422e-a615-194a62af65d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.429722] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278edfac-29f5-42c2-95a8-f822022fd403 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.439212] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e5f5e0-4c90-428e-8fb4-e4dd1433c91d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.452911] env[62235]: DEBUG nova.compute.provider_tree [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.669941] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272220, 'name': Rename_Task, 'duration_secs': 0.159053} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.670270] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.670523] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c274701-6e2b-4eb0-be04-44d50da140f3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.678531] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1100.678531] env[62235]: value = "task-1272221" [ 1100.678531] env[62235]: _type = "Task" [ 1100.678531] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.686855] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.956007] env[62235]: DEBUG nova.scheduler.client.report [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.191224] env[62235]: DEBUG oslo_vmware.api [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272221, 'name': PowerOnVM_Task, 'duration_secs': 0.489997} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.191224] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.191224] env[62235]: INFO nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Took 6.70 seconds to spawn the instance on the hypervisor. [ 1101.191224] env[62235]: DEBUG nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.191546] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d14d87-9ec6-4a51-9013-70b24fade32e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.461642] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.462206] env[62235]: DEBUG nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1101.712412] env[62235]: INFO nova.compute.manager [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Took 11.43 seconds to build instance. [ 1101.967271] env[62235]: DEBUG nova.compute.utils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1101.968724] env[62235]: DEBUG nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1101.968862] env[62235]: DEBUG nova.network.neutron [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1102.010273] env[62235]: DEBUG nova.policy [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '738ee55c831c48f894efbcef71bc20a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c943a445f67f4021bad11723e0836537', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1102.214279] env[62235]: DEBUG oslo_concurrency.lockutils [None req-d463ddf2-62ed-4c5b-bd47-7d8c76a69773 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.942s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.445507] env[62235]: DEBUG nova.network.neutron [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Successfully created port: e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1102.472370] env[62235]: DEBUG nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1102.724166] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "e6421e2f-17cf-4312-a122-067cd032c066" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.724493] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.724694] env[62235]: DEBUG nova.compute.manager [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1102.725985] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd0617b-4ac8-4842-b9a0-ef5f7710c23e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.733510] env[62235]: DEBUG nova.compute.manager [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62235) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1102.734207] env[62235]: DEBUG nova.objects.instance [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'flavor' on Instance uuid e6421e2f-17cf-4312-a122-067cd032c066 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.238981] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.239328] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb1a96ae-638d-4287-8c41-e9103d37d1e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.248031] env[62235]: DEBUG oslo_vmware.api [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1103.248031] env[62235]: value = "task-1272222" [ 1103.248031] env[62235]: _type = "Task" [ 1103.248031] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.256611] env[62235]: DEBUG oslo_vmware.api [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272222, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.484055] env[62235]: DEBUG nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1103.508202] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.508471] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.508695] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.508822] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.508976] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.509197] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.509396] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.509706] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.509768] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.509918] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.510111] env[62235]: DEBUG nova.virt.hardware [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.510978] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a345b1b2-5a04-48bf-8de0-42683f4e2f9f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.519750] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e22d73-cff0-4ca5-8e22-4669047de4ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.758242] env[62235]: DEBUG oslo_vmware.api [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272222, 'name': PowerOffVM_Task, 'duration_secs': 0.209404} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.758610] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.758721] env[62235]: DEBUG nova.compute.manager [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1103.759504] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e822587c-d2cf-4453-9360-f82f333287d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.931808] env[62235]: DEBUG nova.compute.manager [req-1d8a039e-cc84-4113-83f5-6d5acd9f7679 req-88dcdd6f-fe7b-4be6-bb45-db17df56dc86 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Received event network-vif-plugged-e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.932082] env[62235]: DEBUG oslo_concurrency.lockutils [req-1d8a039e-cc84-4113-83f5-6d5acd9f7679 req-88dcdd6f-fe7b-4be6-bb45-db17df56dc86 service nova] Acquiring lock "f7ef033e-ab43-4b89-b331-cb234a373991-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.932308] env[62235]: DEBUG oslo_concurrency.lockutils [req-1d8a039e-cc84-4113-83f5-6d5acd9f7679 req-88dcdd6f-fe7b-4be6-bb45-db17df56dc86 service nova] Lock "f7ef033e-ab43-4b89-b331-cb234a373991-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.932506] env[62235]: DEBUG oslo_concurrency.lockutils [req-1d8a039e-cc84-4113-83f5-6d5acd9f7679 req-88dcdd6f-fe7b-4be6-bb45-db17df56dc86 service nova] Lock "f7ef033e-ab43-4b89-b331-cb234a373991-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.932669] env[62235]: DEBUG nova.compute.manager [req-1d8a039e-cc84-4113-83f5-6d5acd9f7679 req-88dcdd6f-fe7b-4be6-bb45-db17df56dc86 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] No waiting events found dispatching network-vif-plugged-e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1103.932876] env[62235]: WARNING nova.compute.manager [req-1d8a039e-cc84-4113-83f5-6d5acd9f7679 req-88dcdd6f-fe7b-4be6-bb45-db17df56dc86 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Received unexpected event network-vif-plugged-e017afa0-7a52-4f06-8ec1-5e7e24fb337a for instance with vm_state building and task_state spawning. [ 1104.066920] env[62235]: DEBUG nova.network.neutron [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Successfully updated port: e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1104.272624] env[62235]: DEBUG oslo_concurrency.lockutils [None req-09a39c19-a813-48a5-bca3-2d689de3391a tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.573460] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.573460] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.573460] env[62235]: DEBUG nova.network.neutron [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1105.123111] env[62235]: DEBUG nova.network.neutron [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1105.196252] env[62235]: INFO nova.compute.manager [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Rebuilding instance [ 1105.238575] env[62235]: DEBUG nova.compute.manager [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1105.239457] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e17bfd4-74e8-4f55-b8ce-ae5a4d71f524 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.311734] env[62235]: DEBUG nova.network.neutron [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updating instance_info_cache with network_info: [{"id": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "address": "fa:16:3e:c0:27:54", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape017afa0-7a", "ovs_interfaceid": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.751646] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.751959] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a5692c7-cf11-4ee4-b208-cc7fcd8ee988 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.760672] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1105.760672] env[62235]: value = "task-1272223" [ 1105.760672] env[62235]: _type = "Task" [ 1105.760672] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.769032] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272223, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.814743] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.815136] env[62235]: DEBUG nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Instance network_info: |[{"id": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "address": "fa:16:3e:c0:27:54", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape017afa0-7a", "ovs_interfaceid": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1105.815566] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:27:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37fb1918-d178-4e12-93e6-316381e78be4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e017afa0-7a52-4f06-8ec1-5e7e24fb337a', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1105.823385] env[62235]: DEBUG oslo.service.loopingcall [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1105.823522] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1105.823783] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-142ab916-3979-4517-9bb0-6d935f21aa65 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.846487] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1105.846487] env[62235]: value = "task-1272224" [ 1105.846487] env[62235]: _type = "Task" [ 1105.846487] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.856072] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272224, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.971040] env[62235]: DEBUG nova.compute.manager [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Received event network-changed-e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.971040] env[62235]: DEBUG nova.compute.manager [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Refreshing instance network info cache due to event network-changed-e017afa0-7a52-4f06-8ec1-5e7e24fb337a. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1105.971040] env[62235]: DEBUG oslo_concurrency.lockutils [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] Acquiring lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.971040] env[62235]: DEBUG oslo_concurrency.lockutils [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] Acquired lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.971040] env[62235]: DEBUG nova.network.neutron [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Refreshing network info cache for port e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1106.271491] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1106.271834] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1106.272387] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32be4579-8c42-41e9-b732-96fe0434f9fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.288426] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1106.288679] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2874a947-6153-413b-ad6c-f53bb3252704 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.357365] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272224, 'name': CreateVM_Task, 'duration_secs': 0.333864} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.357574] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1106.358282] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.358458] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.358793] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1106.359075] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6070e1c-058d-4a00-ac74-7f7130ed37eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.365228] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1106.365228] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5249b6ee-0753-5345-d1b5-5f02ae26c6bf" [ 1106.365228] env[62235]: _type = "Task" [ 1106.365228] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.370117] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1106.370335] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1106.370773] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleting the datastore file [datastore1] e6421e2f-17cf-4312-a122-067cd032c066 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1106.371234] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce371182-970e-45f6-81c6-424ae3f084d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.376514] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5249b6ee-0753-5345-d1b5-5f02ae26c6bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.378085] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1106.378085] env[62235]: value = "task-1272226" [ 1106.378085] env[62235]: _type = "Task" [ 1106.378085] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.385710] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272226, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.703161] env[62235]: DEBUG nova.network.neutron [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updated VIF entry in instance network info cache for port e017afa0-7a52-4f06-8ec1-5e7e24fb337a. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1106.703535] env[62235]: DEBUG nova.network.neutron [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updating instance_info_cache with network_info: [{"id": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "address": "fa:16:3e:c0:27:54", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape017afa0-7a", "ovs_interfaceid": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.876195] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5249b6ee-0753-5345-d1b5-5f02ae26c6bf, 'name': SearchDatastore_Task, 'duration_secs': 0.011785} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.876486] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.876725] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1106.876963] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.877130] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.877315] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1106.877572] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f8e29f8-b6fb-450c-9a8e-9992d96d328d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.887833] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272226, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12836} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.888961] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1106.889176] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1106.889362] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1106.892039] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1106.892039] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1106.893442] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c0ae047-643d-443a-8f50-ebd21382d2c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.898300] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1106.898300] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523b795e-00aa-967a-ba96-8d8243c52027" [ 1106.898300] env[62235]: _type = "Task" [ 1106.898300] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.906322] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523b795e-00aa-967a-ba96-8d8243c52027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.180428] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.180743] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.180975] env[62235]: INFO nova.compute.manager [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Shelving [ 1107.206351] env[62235]: DEBUG oslo_concurrency.lockutils [req-4d85d3ac-e3f1-4957-9015-95773598228a req-9785ac95-7712-4efa-94bc-99ac3c38f5ae service nova] Releasing lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.409445] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523b795e-00aa-967a-ba96-8d8243c52027, 'name': SearchDatastore_Task, 'duration_secs': 0.009684} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.410285] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d549878f-409b-48f3-b699-2a46ff095b15 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.416759] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1107.416759] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cfe13d-d35b-2528-221c-783b5e2999ee" [ 1107.416759] env[62235]: _type = "Task" [ 1107.416759] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.425228] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cfe13d-d35b-2528-221c-783b5e2999ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.687814] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1107.688033] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb8a4e74-cdec-4ffe-ae2b-4a4100694464 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.696871] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1107.696871] env[62235]: value = "task-1272227" [ 1107.696871] env[62235]: _type = "Task" [ 1107.696871] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.706156] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272227, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.923496] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.923830] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.924029] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.924230] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.924395] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.924586] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.924824] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.925356] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.925356] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.925356] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.925501] env[62235]: DEBUG nova.virt.hardware [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.926274] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e2622e-77fc-4e7c-af0c-3f48516bb0a0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.934666] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cfe13d-d35b-2528-221c-783b5e2999ee, 'name': SearchDatastore_Task, 'duration_secs': 0.010467} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.936835] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.937108] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] f7ef033e-ab43-4b89-b331-cb234a373991/f7ef033e-ab43-4b89-b331-cb234a373991.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1107.937380] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ddfec5d-d730-4a3a-9c66-50630953b843 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.940226] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc745ef0-4f25-49a9-bf80-bc55ed635332 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.956744] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:39:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5357f0fa-5c91-459c-8fc1-4117e140bedb', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1107.963972] env[62235]: DEBUG oslo.service.loopingcall [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.965355] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1107.965765] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1107.965765] env[62235]: value = "task-1272228" [ 1107.965765] env[62235]: _type = "Task" [ 1107.965765] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.966046] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0bbb554-e7a8-4fe3-9cde-637e77892477 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.988245] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.989392] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1107.989392] env[62235]: value = "task-1272229" [ 1107.989392] env[62235]: _type = "Task" [ 1107.989392] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.997317] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272229, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.139151] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.139482] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.210468] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272227, 'name': PowerOffVM_Task, 'duration_secs': 0.171531} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.210773] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1108.211616] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f26a81-0d45-4c91-a79c-e2fd9589cc52 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.236167] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6d57c2-abbd-40ed-a154-32737cd45610 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.489483] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272228, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459557} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.489843] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] f7ef033e-ab43-4b89-b331-cb234a373991/f7ef033e-ab43-4b89-b331-cb234a373991.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1108.490060] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1108.492884] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63e6c3f1-f13f-483d-8404-e0db94622d5a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.499429] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272229, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.500541] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1108.500541] env[62235]: value = "task-1272230" [ 1108.500541] env[62235]: _type = "Task" [ 1108.500541] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.507397] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272230, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.643281] env[62235]: DEBUG nova.compute.utils [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1108.749836] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Creating Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1108.750269] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bce56b24-a588-45f4-b8e9-6bcee6a818b0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.759947] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1108.759947] env[62235]: value = "task-1272231" [ 1108.759947] env[62235]: _type = "Task" [ 1108.759947] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.770278] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272231, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.001937] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272229, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.009947] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272230, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.145836] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.270827] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272231, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.500731] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272229, 'name': CreateVM_Task, 'duration_secs': 1.366976} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.501094] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1109.501566] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.501739] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.502081] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.502391] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dcf1778-7c80-45d1-8e89-3c8cb5a8f16c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.509497] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1109.509497] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528fc2fc-8eda-48d8-f4a7-c214aac680af" [ 1109.509497] env[62235]: _type = "Task" [ 1109.509497] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.512663] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272230, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.825247} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.515377] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1109.516120] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcc04a9-7e5a-46c8-bea1-68250d7fd6d0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.540299] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] f7ef033e-ab43-4b89-b331-cb234a373991/f7ef033e-ab43-4b89-b331-cb234a373991.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.540609] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]528fc2fc-8eda-48d8-f4a7-c214aac680af, 'name': SearchDatastore_Task, 'duration_secs': 0.010138} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.540823] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-599f93b0-5959-449e-a8c3-d7d9f0f07518 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.554406] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.555063] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1109.555063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.555063] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.555280] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1109.555435] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-137359c7-990e-476e-a277-eecab060fc98 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.564126] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1109.564126] env[62235]: value = "task-1272232" [ 1109.564126] env[62235]: _type = "Task" [ 1109.564126] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.565154] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1109.565335] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1109.568460] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9c2f45f-1ada-40a7-a363-9549529e915b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.573412] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1109.573412] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a06375-5f1d-ff77-459d-7c0a6a1e624e" [ 1109.573412] env[62235]: _type = "Task" [ 1109.573412] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.576766] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272232, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.584677] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a06375-5f1d-ff77-459d-7c0a6a1e624e, 'name': SearchDatastore_Task, 'duration_secs': 0.00899} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.585424] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5cb916e-7ed5-4c49-8f5f-3df84f4adf42 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.590468] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1109.590468] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522c23ed-d94c-82fa-150e-f74c3f72f418" [ 1109.590468] env[62235]: _type = "Task" [ 1109.590468] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.597717] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522c23ed-d94c-82fa-150e-f74c3f72f418, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.770939] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272231, 'name': CreateSnapshot_Task, 'duration_secs': 0.919603} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.771278] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Created Snapshot of the VM instance {{(pid=62235) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1109.772082] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8224051-4a82-499e-8ae1-79f454658ed8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.075159] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272232, 'name': ReconfigVM_Task, 'duration_secs': 0.281106} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.075509] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Reconfigured VM instance instance-00000069 to attach disk [datastore2] f7ef033e-ab43-4b89-b331-cb234a373991/f7ef033e-ab43-4b89-b331-cb234a373991.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.076204] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e58d0f4d-e10b-4ca1-ab2b-65b9351f15dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.084086] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1110.084086] env[62235]: value = "task-1272233" [ 1110.084086] env[62235]: _type = "Task" [ 1110.084086] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.093602] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272233, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.101361] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522c23ed-d94c-82fa-150e-f74c3f72f418, 'name': SearchDatastore_Task, 'duration_secs': 0.007926} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.101599] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.101852] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1110.102106] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4932d8e-0919-46bb-bd65-dbbcd7b86911 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.109211] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1110.109211] env[62235]: value = "task-1272234" [ 1110.109211] env[62235]: _type = "Task" [ 1110.109211] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.117039] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.293373] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Creating linked-clone VM from snapshot {{(pid=62235) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1110.294695] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.294695] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.294695] env[62235]: INFO nova.compute.manager [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Attaching volume d47a6600-d7c3-4edf-860f-557cb990e752 to /dev/sdb [ 1110.296019] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3c11b6bc-ac73-4d10-9a43-508bc1deafc7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.306577] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1110.306577] env[62235]: value = "task-1272235" [ 1110.306577] env[62235]: _type = "Task" [ 1110.306577] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.316780] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272235, 'name': CloneVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.337575] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a9204e-effe-4c27-9d6b-398bd94565f0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.345879] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b784f717-85ec-4e60-bb69-bd7dfd7391ca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.362238] env[62235]: DEBUG nova.virt.block_device [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updating existing volume attachment record: f0dae66c-4477-46d9-b2c9-c6d50a9554ca {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1110.596341] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272233, 'name': Rename_Task, 'duration_secs': 0.1449} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.596767] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1110.597060] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a8b1469-2125-4b98-bfa9-7c6b4358057b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.605935] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1110.605935] env[62235]: value = "task-1272237" [ 1110.605935] env[62235]: _type = "Task" [ 1110.605935] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.617036] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.623974] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272234, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481908} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.624319] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1110.624547] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1110.624857] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da46bcc1-adf1-4946-b742-75f6d14961de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.632836] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1110.632836] env[62235]: value = "task-1272238" [ 1110.632836] env[62235]: _type = "Task" [ 1110.632836] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.652051] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272238, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.818109] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272235, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.115954] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272237, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.141589] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272238, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069014} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.141896] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1111.142719] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e46c44-b38c-452b-bb1f-4c4940181f4c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.164827] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.165184] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80341457-7208-4c78-b713-672f8c6f03ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.185788] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1111.185788] env[62235]: value = "task-1272239" [ 1111.185788] env[62235]: _type = "Task" [ 1111.185788] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.194067] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.320269] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272235, 'name': CloneVM_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.615864] env[62235]: DEBUG oslo_vmware.api [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272237, 'name': PowerOnVM_Task, 'duration_secs': 0.822459} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.616221] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1111.616439] env[62235]: INFO nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Took 8.13 seconds to spawn the instance on the hypervisor. [ 1111.616626] env[62235]: DEBUG nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1111.617374] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36afc60d-9185-4bf1-b829-8965bc6da510 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.697188] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272239, 'name': ReconfigVM_Task, 'duration_secs': 0.267943} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.697832] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Reconfigured VM instance instance-00000068 to attach disk [datastore1] e6421e2f-17cf-4312-a122-067cd032c066/e6421e2f-17cf-4312-a122-067cd032c066.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1111.697978] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d09f893-24f3-4fe1-b99b-1a00895b76da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.707141] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1111.707141] env[62235]: value = "task-1272240" [ 1111.707141] env[62235]: _type = "Task" [ 1111.707141] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.716454] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272240, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.818957] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272235, 'name': CloneVM_Task} progress is 95%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.135877] env[62235]: INFO nova.compute.manager [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Took 12.86 seconds to build instance. [ 1112.220790] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272240, 'name': Rename_Task, 'duration_secs': 0.145353} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.220790] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1112.220790] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f87370b-3321-4d1b-9061-7fb7d52344be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.228083] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1112.228083] env[62235]: value = "task-1272241" [ 1112.228083] env[62235]: _type = "Task" [ 1112.228083] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.235591] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272241, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.319267] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272235, 'name': CloneVM_Task, 'duration_secs': 1.779475} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.319561] env[62235]: INFO nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Created linked-clone VM from snapshot [ 1112.320329] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c640958b-32bf-4d4a-a96b-5619b0acb541 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.330765] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Uploading image 23fc41d5-108d-47df-b47c-d7f47435d593 {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1112.356289] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1112.356289] env[62235]: value = "vm-273528" [ 1112.356289] env[62235]: _type = "VirtualMachine" [ 1112.356289] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1112.356577] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3dcedfb7-ba7f-4d69-b81a-6d3616941f15 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.364174] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lease: (returnval){ [ 1112.364174] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e3e0ca-3944-87ac-e8fd-058c83e600d5" [ 1112.364174] env[62235]: _type = "HttpNfcLease" [ 1112.364174] env[62235]: } obtained for exporting VM: (result){ [ 1112.364174] env[62235]: value = "vm-273528" [ 1112.364174] env[62235]: _type = "VirtualMachine" [ 1112.364174] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1112.364426] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the lease: (returnval){ [ 1112.364426] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e3e0ca-3944-87ac-e8fd-058c83e600d5" [ 1112.364426] env[62235]: _type = "HttpNfcLease" [ 1112.364426] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1112.371062] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1112.371062] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e3e0ca-3944-87ac-e8fd-058c83e600d5" [ 1112.371062] env[62235]: _type = "HttpNfcLease" [ 1112.371062] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1112.485248] env[62235]: DEBUG nova.compute.manager [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Received event network-changed-e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.485456] env[62235]: DEBUG nova.compute.manager [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Refreshing instance network info cache due to event network-changed-e017afa0-7a52-4f06-8ec1-5e7e24fb337a. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1112.485682] env[62235]: DEBUG oslo_concurrency.lockutils [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] Acquiring lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.485870] env[62235]: DEBUG oslo_concurrency.lockutils [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] Acquired lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.486362] env[62235]: DEBUG nova.network.neutron [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Refreshing network info cache for port e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.639261] env[62235]: DEBUG oslo_concurrency.lockutils [None req-db3730be-d455-44ca-ad23-b52ac2076d83 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.375s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.737943] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272241, 'name': PowerOnVM_Task, 'duration_secs': 0.436479} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.738363] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1112.738654] env[62235]: DEBUG nova.compute.manager [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1112.739569] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc06340-ce71-416e-bdd9-b7e51fd305e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.874114] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1112.874114] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e3e0ca-3944-87ac-e8fd-058c83e600d5" [ 1112.874114] env[62235]: _type = "HttpNfcLease" [ 1112.874114] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1112.874407] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1112.874407] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e3e0ca-3944-87ac-e8fd-058c83e600d5" [ 1112.874407] env[62235]: _type = "HttpNfcLease" [ 1112.874407] env[62235]: }. {{(pid=62235) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1112.875161] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0eb8b3f-b030-4820-9670-913eb49c4a9a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.882579] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c359-a18e-a450-62ef-1ffd56c6f299/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1112.882768] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c359-a18e-a450-62ef-1ffd56c6f299/disk-0.vmdk for reading. {{(pid=62235) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1112.977475] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-06907743-1d24-4ad3-be58-a80cf025b1c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.221491] env[62235]: DEBUG nova.network.neutron [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updated VIF entry in instance network info cache for port e017afa0-7a52-4f06-8ec1-5e7e24fb337a. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1113.222101] env[62235]: DEBUG nova.network.neutron [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updating instance_info_cache with network_info: [{"id": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "address": "fa:16:3e:c0:27:54", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape017afa0-7a", "ovs_interfaceid": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.251431] env[62235]: INFO nova.compute.manager [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] bringing vm to original state: 'stopped' [ 1113.724864] env[62235]: DEBUG oslo_concurrency.lockutils [req-9e66a1a9-4f8f-4a49-9705-0845b6faf2f6 req-b722548b-78d2-460b-bad0-7c2d7c752446 service nova] Releasing lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.259524] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "e6421e2f-17cf-4312-a122-067cd032c066" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.259524] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.259875] env[62235]: DEBUG nova.compute.manager [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.260941] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00bb74d3-de3a-4906-bc4c-104bfff79c1c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.268622] env[62235]: DEBUG nova.compute.manager [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62235) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1114.270686] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.270998] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ebc8122-7a63-4a9e-8ec9-9ad6379b17f7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.287191] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1114.287191] env[62235]: value = "task-1272244" [ 1114.287191] env[62235]: _type = "Task" [ 1114.287191] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.296463] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.797614] env[62235]: DEBUG oslo_vmware.api [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272244, 'name': PowerOffVM_Task, 'duration_secs': 0.286048} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.798050] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.798289] env[62235]: DEBUG nova.compute.manager [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.799108] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e22c72-2da6-4ad5-945b-b6702596544b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.912710] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1114.913049] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273529', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'name': 'volume-d47a6600-d7c3-4edf-860f-557cb990e752', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a993bed1-e346-47e7-9f7f-abdfb6685196', 'attached_at': '', 'detached_at': '', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'serial': 'd47a6600-d7c3-4edf-860f-557cb990e752'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1114.914139] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a78c156-b049-425b-9e7f-a45c555f9db6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.935214] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66549ed9-b36b-4b01-934f-a7f35ec69f3a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.963035] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] volume-d47a6600-d7c3-4edf-860f-557cb990e752/volume-d47a6600-d7c3-4edf-860f-557cb990e752.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.963402] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86c8be69-5bd1-46c8-98fd-006eed0436e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.984506] env[62235]: DEBUG oslo_vmware.api [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1114.984506] env[62235]: value = "task-1272245" [ 1114.984506] env[62235]: _type = "Task" [ 1114.984506] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.995737] env[62235]: DEBUG oslo_vmware.api [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272245, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.312319] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.053s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.494604] env[62235]: DEBUG oslo_vmware.api [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272245, 'name': ReconfigVM_Task, 'duration_secs': 0.388908} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.494900] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Reconfigured VM instance instance-00000065 to attach disk [datastore1] volume-d47a6600-d7c3-4edf-860f-557cb990e752/volume-d47a6600-d7c3-4edf-860f-557cb990e752.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.499959] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2b3d9c6-5339-44c4-bf87-df84e67ad398 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.515616] env[62235]: DEBUG oslo_vmware.api [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1115.515616] env[62235]: value = "task-1272246" [ 1115.515616] env[62235]: _type = "Task" [ 1115.515616] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.523641] env[62235]: DEBUG oslo_vmware.api [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272246, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.823503] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.823760] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.823995] env[62235]: DEBUG nova.objects.instance [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1116.027522] env[62235]: DEBUG oslo_vmware.api [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272246, 'name': ReconfigVM_Task, 'duration_secs': 0.139944} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.027991] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273529', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'name': 'volume-d47a6600-d7c3-4edf-860f-557cb990e752', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a993bed1-e346-47e7-9f7f-abdfb6685196', 'attached_at': '', 'detached_at': '', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'serial': 'd47a6600-d7c3-4edf-860f-557cb990e752'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1116.798553] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "e6421e2f-17cf-4312-a122-067cd032c066" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.798864] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.799156] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "e6421e2f-17cf-4312-a122-067cd032c066-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.799382] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.799621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.801969] env[62235]: INFO nova.compute.manager [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Terminating instance [ 1116.804342] env[62235]: DEBUG nova.compute.manager [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1116.804574] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1116.805512] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b1031e-6cc1-42c7-9402-f8d4675c74d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.814439] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1116.814703] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e81e4747-2f0f-4b93-b9ff-0220f37ee389 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.831682] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4e47273d-0535-4adf-a618-7a791e5980cd tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.895072] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1116.895297] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1116.895492] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleting the datastore file [datastore1] e6421e2f-17cf-4312-a122-067cd032c066 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.895839] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a0d7d1f-760e-4f15-8e17-2285dc404ff6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.903889] env[62235]: DEBUG oslo_vmware.api [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1116.903889] env[62235]: value = "task-1272248" [ 1116.903889] env[62235]: _type = "Task" [ 1116.903889] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.913992] env[62235]: DEBUG oslo_vmware.api [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272248, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.100115] env[62235]: DEBUG nova.objects.instance [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'flavor' on Instance uuid a993bed1-e346-47e7-9f7f-abdfb6685196 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.287219] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.287530] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.287785] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.288047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.288292] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.290620] env[62235]: INFO nova.compute.manager [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Terminating instance [ 1117.292637] env[62235]: DEBUG nova.compute.manager [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1117.292879] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.293847] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472a111f-0a22-4502-98e4-7e892c1597cb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.303035] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.303287] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea1f15aa-e92f-4e7b-9e9c-da3b35122021 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.309901] env[62235]: DEBUG oslo_vmware.api [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1117.309901] env[62235]: value = "task-1272249" [ 1117.309901] env[62235]: _type = "Task" [ 1117.309901] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.317629] env[62235]: DEBUG oslo_vmware.api [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272249, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.414792] env[62235]: DEBUG oslo_vmware.api [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272248, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179852} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.415068] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.415278] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1117.415465] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1117.415653] env[62235]: INFO nova.compute.manager [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1117.415907] env[62235]: DEBUG oslo.service.loopingcall [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.416126] env[62235]: DEBUG nova.compute.manager [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1117.416224] env[62235]: DEBUG nova.network.neutron [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1117.605685] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75baceb7-bf25-4174-bfff-25eb6afcb4fa tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.311s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.688609] env[62235]: DEBUG nova.compute.manager [req-215ff980-ba61-4dc5-a9d7-024ff651a5b5 req-5629811f-90ce-4f4e-9409-80204dd3e6b6 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Received event network-vif-deleted-5357f0fa-5c91-459c-8fc1-4117e140bedb {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1117.688795] env[62235]: INFO nova.compute.manager [req-215ff980-ba61-4dc5-a9d7-024ff651a5b5 req-5629811f-90ce-4f4e-9409-80204dd3e6b6 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Neutron deleted interface 5357f0fa-5c91-459c-8fc1-4117e140bedb; detaching it from the instance and deleting it from the info cache [ 1117.688974] env[62235]: DEBUG nova.network.neutron [req-215ff980-ba61-4dc5-a9d7-024ff651a5b5 req-5629811f-90ce-4f4e-9409-80204dd3e6b6 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.735069] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.735069] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.819530] env[62235]: DEBUG oslo_vmware.api [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272249, 'name': PowerOffVM_Task, 'duration_secs': 0.25521} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.819898] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1117.820128] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1117.820432] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fae38f9-b53d-456a-a198-6ec9961e2da1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.085987] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.085987] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.086361] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleting the datastore file [datastore2] 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.086475] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da239467-4d45-41c7-b728-9b2e4846d8ea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.094521] env[62235]: DEBUG oslo_vmware.api [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for the task: (returnval){ [ 1118.094521] env[62235]: value = "task-1272251" [ 1118.094521] env[62235]: _type = "Task" [ 1118.094521] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.102946] env[62235]: DEBUG oslo_vmware.api [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.165753] env[62235]: DEBUG nova.network.neutron [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.191666] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39b3e31c-a1ee-4f46-b7fe-8bce210edbfc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.202332] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d293c771-8113-4af3-a05d-52c82e072d6d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.231078] env[62235]: DEBUG nova.compute.manager [req-215ff980-ba61-4dc5-a9d7-024ff651a5b5 req-5629811f-90ce-4f4e-9409-80204dd3e6b6 service nova] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Detach interface failed, port_id=5357f0fa-5c91-459c-8fc1-4117e140bedb, reason: Instance e6421e2f-17cf-4312-a122-067cd032c066 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1118.238034] env[62235]: INFO nova.compute.manager [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Detaching volume d47a6600-d7c3-4edf-860f-557cb990e752 [ 1118.275257] env[62235]: INFO nova.virt.block_device [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Attempting to driver detach volume d47a6600-d7c3-4edf-860f-557cb990e752 from mountpoint /dev/sdb [ 1118.275507] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1118.275704] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273529', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'name': 'volume-d47a6600-d7c3-4edf-860f-557cb990e752', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a993bed1-e346-47e7-9f7f-abdfb6685196', 'attached_at': '', 'detached_at': '', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'serial': 'd47a6600-d7c3-4edf-860f-557cb990e752'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1118.276640] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35dc88a5-2e8b-47e7-9315-0cbb81528232 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.303287] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c599a5-c3ce-4658-867b-294eb61a570d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.310919] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec4a076-072a-417a-9613-6d5132fea0c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.331631] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70387afe-bef3-4ae0-814c-2c8e32467fa1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.347053] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] The volume has not been displaced from its original location: [datastore1] volume-d47a6600-d7c3-4edf-860f-557cb990e752/volume-d47a6600-d7c3-4edf-860f-557cb990e752.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1118.352589] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1118.352943] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-371b2b46-e3fb-4ae4-8f2f-68674be26a09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.372348] env[62235]: DEBUG oslo_vmware.api [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1118.372348] env[62235]: value = "task-1272252" [ 1118.372348] env[62235]: _type = "Task" [ 1118.372348] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.380371] env[62235]: DEBUG oslo_vmware.api [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272252, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.604255] env[62235]: DEBUG oslo_vmware.api [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Task: {'id': task-1272251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.436678} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.604501] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1118.604676] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1118.604861] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1118.605048] env[62235]: INFO nova.compute.manager [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Took 1.31 seconds to destroy the instance on the hypervisor. [ 1118.605309] env[62235]: DEBUG oslo.service.loopingcall [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1118.605529] env[62235]: DEBUG nova.compute.manager [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1118.605627] env[62235]: DEBUG nova.network.neutron [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1118.668376] env[62235]: INFO nova.compute.manager [-] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Took 1.25 seconds to deallocate network for instance. [ 1118.871635] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.872058] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.884438] env[62235]: DEBUG oslo_vmware.api [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272252, 'name': ReconfigVM_Task, 'duration_secs': 0.266415} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.884762] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1118.891107] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5b788c1-2df7-4996-8489-fc3d30373dee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.911066] env[62235]: DEBUG oslo_vmware.api [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1118.911066] env[62235]: value = "task-1272253" [ 1118.911066] env[62235]: _type = "Task" [ 1118.911066] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.923812] env[62235]: DEBUG oslo_vmware.api [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272253, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.175678] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.176011] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.176288] env[62235]: DEBUG nova.objects.instance [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'resources' on Instance uuid e6421e2f-17cf-4312-a122-067cd032c066 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.382846] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.383086] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1119.420967] env[62235]: DEBUG oslo_vmware.api [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272253, 'name': ReconfigVM_Task, 'duration_secs': 0.163003} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.421107] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273529', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'name': 'volume-d47a6600-d7c3-4edf-860f-557cb990e752', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a993bed1-e346-47e7-9f7f-abdfb6685196', 'attached_at': '', 'detached_at': '', 'volume_id': 'd47a6600-d7c3-4edf-860f-557cb990e752', 'serial': 'd47a6600-d7c3-4edf-860f-557cb990e752'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1119.554914] env[62235]: DEBUG nova.network.neutron [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.714471] env[62235]: DEBUG nova.compute.manager [req-9636ec7a-cdb1-4427-9c80-51461193761f req-40b8d4e3-97f1-47a5-97ce-41acc3a1805a service nova] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Received event network-vif-deleted-808e6de1-f745-433c-b5e8-705d1fb7466b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.778969] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06043de-bc02-4b77-9216-bea539a483d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.787098] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76376b5e-d0b3-44fa-8b44-6bc079a86578 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.819398] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6ae9f5-72f6-4943-b84b-89766add7283 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.828245] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1559fc3d-735d-4db5-bd5a-1000f7d53c23 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.844632] env[62235]: DEBUG nova.compute.provider_tree [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.969650] env[62235]: DEBUG nova.objects.instance [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'flavor' on Instance uuid a993bed1-e346-47e7-9f7f-abdfb6685196 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.058289] env[62235]: INFO nova.compute.manager [-] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Took 1.45 seconds to deallocate network for instance. [ 1120.347626] env[62235]: DEBUG nova.scheduler.client.report [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.445547] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.445711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.445935] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1120.565290] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.852691] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.855162] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.290s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.855406] env[62235]: DEBUG nova.objects.instance [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lazy-loading 'resources' on Instance uuid 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.883030] env[62235]: INFO nova.scheduler.client.report [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted allocations for instance e6421e2f-17cf-4312-a122-067cd032c066 [ 1120.978292] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b2c90853-b21b-4c38-ab80-1dc1506ca742 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.243s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.393140] env[62235]: DEBUG oslo_concurrency.lockutils [None req-905ba08f-2169-47a1-a02d-aef665c91d8d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "e6421e2f-17cf-4312-a122-067cd032c066" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.594s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.439848] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312bee8d-e819-4783-b9fd-dd3d813232c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.448059] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc098723-9869-430c-b248-538908eb0168 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.480063] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8154b4-86dc-46b7-b7cf-59ab0bd5ee7e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.489907] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32f0e69-ad2d-4f84-9304-e9ba39a5d99c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.503984] env[62235]: DEBUG nova.compute.provider_tree [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.708796] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.009937] env[62235]: DEBUG nova.scheduler.client.report [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.013361] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.013650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.013863] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.014563] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.014563] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.016473] env[62235]: INFO nova.compute.manager [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Terminating instance [ 1122.017880] env[62235]: DEBUG nova.compute.manager [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1122.018084] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1122.019055] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008e4c35-2009-40d8-a6c8-846dba0ef970 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.027715] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1122.028558] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7196f2a-8d4a-418f-aa15-e0111264c012 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.035762] env[62235]: DEBUG oslo_vmware.api [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1122.035762] env[62235]: value = "task-1272254" [ 1122.035762] env[62235]: _type = "Task" [ 1122.035762] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.044801] env[62235]: DEBUG oslo_vmware.api [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272254, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.211517] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.211803] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1122.212072] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.212272] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.212497] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.212683] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.212845] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.212995] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.460906] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c359-a18e-a450-62ef-1ffd56c6f299/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1122.461889] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b1a9fb-996a-4a03-aa29-965a5732ced9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.468353] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c359-a18e-a450-62ef-1ffd56c6f299/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1122.468527] env[62235]: ERROR oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c359-a18e-a450-62ef-1ffd56c6f299/disk-0.vmdk due to incomplete transfer. [ 1122.468764] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-96657af5-8741-4cce-b802-d68f4588d962 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.476033] env[62235]: DEBUG oslo_vmware.rw_handles [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f0c359-a18e-a450-62ef-1ffd56c6f299/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1122.476245] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Uploaded image 23fc41d5-108d-47df-b47c-d7f47435d593 to the Glance image server {{(pid=62235) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1122.478584] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Destroying the VM {{(pid=62235) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1122.478844] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fd17da1b-2a8d-480f-8a71-2f0ca8b7bf4f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.484873] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1122.484873] env[62235]: value = "task-1272255" [ 1122.484873] env[62235]: _type = "Task" [ 1122.484873] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.492955] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272255, 'name': Destroy_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.517059] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.662s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.546640] env[62235]: DEBUG oslo_vmware.api [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272254, 'name': PowerOffVM_Task, 'duration_secs': 0.194075} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.547499] env[62235]: INFO nova.scheduler.client.report [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Deleted allocations for instance 738750ad-931e-45e4-b0c6-8232fe1b9f46 [ 1122.548507] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1122.551167] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1122.553680] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3b7e088-f16e-4701-92a0-2454a054af5a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.717106] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Getting list of instances from cluster (obj){ [ 1122.717106] env[62235]: value = "domain-c8" [ 1122.717106] env[62235]: _type = "ClusterComputeResource" [ 1122.717106] env[62235]: } {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1122.718603] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc10d09-ddee-4efb-9e12-bc5397628d7d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.733751] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Got total of 4 instances {{(pid=62235) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1122.733909] env[62235]: WARNING nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] While synchronizing instance power states, found 5 instances in the database and 4 instances on the hypervisor. [ 1122.734056] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Triggering sync for uuid 738750ad-931e-45e4-b0c6-8232fe1b9f46 {{(pid=62235) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1122.734252] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Triggering sync for uuid 0feb5395-f648-419a-998b-eff941c16464 {{(pid=62235) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1122.734405] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Triggering sync for uuid a993bed1-e346-47e7-9f7f-abdfb6685196 {{(pid=62235) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1122.734555] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Triggering sync for uuid 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1122.734704] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Triggering sync for uuid f7ef033e-ab43-4b89-b331-cb234a373991 {{(pid=62235) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1122.735437] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.735437] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "0feb5395-f648-419a-998b-eff941c16464" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.735698] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "0feb5395-f648-419a-998b-eff941c16464" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.735757] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "a993bed1-e346-47e7-9f7f-abdfb6685196" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.735924] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.736209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "f7ef033e-ab43-4b89-b331-cb234a373991" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.736322] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.736525] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.736680] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1122.737361] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c504f43-2e0f-48fe-a50b-c8cf86c6e30c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.740348] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab75193-ee01-444a-bb4e-e7d34d3906c1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.742820] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.832279] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1122.832563] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1122.832758] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleting the datastore file [datastore2] a993bed1-e346-47e7-9f7f-abdfb6685196 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1122.833341] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d12168c-c0f2-4137-a91b-8e70b3e0ba79 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.843580] env[62235]: DEBUG oslo_vmware.api [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for the task: (returnval){ [ 1122.843580] env[62235]: value = "task-1272257" [ 1122.843580] env[62235]: _type = "Task" [ 1122.843580] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.851613] env[62235]: DEBUG oslo_vmware.api [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.994902] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272255, 'name': Destroy_Task, 'duration_secs': 0.445863} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.995214] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Destroyed the VM [ 1122.995503] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Deleting Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1122.995767] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-19e25535-0606-436e-a092-acdbda4aeae7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.003576] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1123.003576] env[62235]: value = "task-1272258" [ 1123.003576] env[62235]: _type = "Task" [ 1123.003576] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.011604] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272258, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.059477] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f371849-2363-44d4-a6bb-21200f5bde67 tempest-AttachVolumeShelveTestJSON-347299526 tempest-AttachVolumeShelveTestJSON-347299526-project-member] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.772s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.060870] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.325s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.060870] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e64c33b2-12d8-4f45-8918-e44e3f9d32e8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.069736] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e94deef-5865-4c4b-b53e-73be6895a7bd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.245859] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.246116] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.246296] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.246492] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1123.247411] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4322447-9d2d-4488-915f-bea9ab11cd73 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.252095] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "0feb5395-f648-419a-998b-eff941c16464" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.252472] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.255731] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2ccb7a-458e-4c0d-afa2-7e985133ee75 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.269280] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76393a26-2a72-4369-99f6-80802750fd8f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.275638] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868f7b9e-f20f-4bd0-a0ae-d389fc8d13fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.304730] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181001MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1123.304730] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.304936] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.354166] env[62235]: DEBUG oslo_vmware.api [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Task: {'id': task-1272257, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162811} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.354438] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1123.354634] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1123.354819] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1123.354996] env[62235]: INFO nova.compute.manager [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Took 1.34 seconds to destroy the instance on the hypervisor. [ 1123.355263] env[62235]: DEBUG oslo.service.loopingcall [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.355464] env[62235]: DEBUG nova.compute.manager [-] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1123.355576] env[62235]: DEBUG nova.network.neutron [-] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1123.514075] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272258, 'name': RemoveSnapshot_Task, 'duration_secs': 0.363868} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.514075] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Deleted Snapshot of the VM instance {{(pid=62235) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1123.514382] env[62235]: DEBUG nova.compute.manager [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1123.515306] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed537163-016e-4698-9543-cde8742e37b0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.599424] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "738750ad-931e-45e4-b0c6-8232fe1b9f46" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.539s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.899940] env[62235]: DEBUG nova.compute.manager [req-91038472-538a-4902-b87a-fb45551807b9 req-c3084bb6-2c06-47da-a969-e886c8ec94af service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Received event network-vif-deleted-c973dab3-9201-4fb9-8407-9e0eb19895d4 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1123.900183] env[62235]: INFO nova.compute.manager [req-91038472-538a-4902-b87a-fb45551807b9 req-c3084bb6-2c06-47da-a969-e886c8ec94af service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Neutron deleted interface c973dab3-9201-4fb9-8407-9e0eb19895d4; detaching it from the instance and deleting it from the info cache [ 1123.900504] env[62235]: DEBUG nova.network.neutron [req-91038472-538a-4902-b87a-fb45551807b9 req-c3084bb6-2c06-47da-a969-e886c8ec94af service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.027743] env[62235]: INFO nova.compute.manager [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Shelve offloading [ 1124.029408] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1124.029665] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4dbec229-e881-4f3a-a3aa-6756e9a0763f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.037881] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1124.037881] env[62235]: value = "task-1272259" [ 1124.037881] env[62235]: _type = "Task" [ 1124.037881] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.046358] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272259, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.195819] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.196090] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.331463] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0feb5395-f648-419a-998b-eff941c16464 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.331463] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a993bed1-e346-47e7-9f7f-abdfb6685196 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.331463] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 5e52ce21-04a1-41db-b62e-2932bf91538e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.331645] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance f7ef033e-ab43-4b89-b331-cb234a373991 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.378138] env[62235]: DEBUG nova.network.neutron [-] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.402943] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5934ef50-995e-4db0-a0e5-51f0a06fab40 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.412589] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6610187-eb54-45a1-85d0-f7d100662624 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.438138] env[62235]: DEBUG nova.compute.manager [req-91038472-538a-4902-b87a-fb45551807b9 req-c3084bb6-2c06-47da-a969-e886c8ec94af service nova] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Detach interface failed, port_id=c973dab3-9201-4fb9-8407-9e0eb19895d4, reason: Instance a993bed1-e346-47e7-9f7f-abdfb6685196 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1124.548414] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1124.548757] env[62235]: DEBUG nova.compute.manager [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.549402] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75557aa7-1cb0-4c6c-8f85-6b605814fd8e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.554854] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.555034] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.555212] env[62235]: DEBUG nova.network.neutron [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1124.698804] env[62235]: DEBUG nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1124.834808] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a293951c-183d-4bfe-b41e-2b3408f2eee3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1124.835075] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1124.835238] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1124.880788] env[62235]: INFO nova.compute.manager [-] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Took 1.53 seconds to deallocate network for instance. [ 1124.909162] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915ec5f9-393e-4d81-88c9-97afbd829905 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.916658] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e938fdd-a6fe-4a72-bb87-c78d3f33cc33 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.949430] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afec9199-cd34-40b9-9d25-00152f4eec6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.957042] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14870781-e72c-4438-9280-54f0730dd362 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.971420] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.219454] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.275663] env[62235]: DEBUG nova.network.neutron [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b639a75-67", "ovs_interfaceid": "3b639a75-6787-4039-95d9-4197fe7bff26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.388832] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.474370] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.778504] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.979249] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1125.979486] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.675s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.979779] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.760s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.981532] env[62235]: INFO nova.compute.claims [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1126.032987] env[62235]: DEBUG nova.compute.manager [req-e1937094-cf9a-4259-9832-06e43535f8da req-a1c4742a-a72a-49ff-9463-93eed0eb7f0d service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received event network-vif-unplugged-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1126.032987] env[62235]: DEBUG oslo_concurrency.lockutils [req-e1937094-cf9a-4259-9832-06e43535f8da req-a1c4742a-a72a-49ff-9463-93eed0eb7f0d service nova] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.033300] env[62235]: DEBUG oslo_concurrency.lockutils [req-e1937094-cf9a-4259-9832-06e43535f8da req-a1c4742a-a72a-49ff-9463-93eed0eb7f0d service nova] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.033502] env[62235]: DEBUG oslo_concurrency.lockutils [req-e1937094-cf9a-4259-9832-06e43535f8da req-a1c4742a-a72a-49ff-9463-93eed0eb7f0d service nova] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.033724] env[62235]: DEBUG nova.compute.manager [req-e1937094-cf9a-4259-9832-06e43535f8da req-a1c4742a-a72a-49ff-9463-93eed0eb7f0d service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] No waiting events found dispatching network-vif-unplugged-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1126.033929] env[62235]: WARNING nova.compute.manager [req-e1937094-cf9a-4259-9832-06e43535f8da req-a1c4742a-a72a-49ff-9463-93eed0eb7f0d service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received unexpected event network-vif-unplugged-3b639a75-6787-4039-95d9-4197fe7bff26 for instance with vm_state shelved and task_state shelving_offloading. [ 1126.056584] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1126.057912] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643815cb-a348-4281-be89-3ec9c35704bd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.067438] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1126.067705] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb1634b6-c269-4795-ad40-67d84533cc4d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.137054] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1126.137314] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1126.137509] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleting the datastore file [datastore2] 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1126.137805] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-245bdb9c-9688-4248-908c-10e9004f21e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.145488] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1126.145488] env[62235]: value = "task-1272262" [ 1126.145488] env[62235]: _type = "Task" [ 1126.145488] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.153031] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272262, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.655595] env[62235]: DEBUG oslo_vmware.api [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272262, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120271} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.655790] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.655980] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1126.656183] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1126.674163] env[62235]: INFO nova.scheduler.client.report [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleted allocations for instance 5e52ce21-04a1-41db-b62e-2932bf91538e [ 1127.051557] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959122b3-f679-4035-b518-9b421bfc5a6d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.062245] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f1f40c-5415-45c4-9b59-6494bf8f6d23 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.092845] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20933918-e72b-4062-a2a3-c93813a03be2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.101642] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629fd03f-b88e-4d66-80c8-9b2c6a3b2f13 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.114609] env[62235]: DEBUG nova.compute.provider_tree [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.179010] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.617646] env[62235]: DEBUG nova.scheduler.client.report [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.084083] env[62235]: DEBUG nova.compute.manager [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received event network-changed-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.084448] env[62235]: DEBUG nova.compute.manager [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Refreshing instance network info cache due to event network-changed-3b639a75-6787-4039-95d9-4197fe7bff26. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1128.084537] env[62235]: DEBUG oslo_concurrency.lockutils [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] Acquiring lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.084621] env[62235]: DEBUG oslo_concurrency.lockutils [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] Acquired lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.084835] env[62235]: DEBUG nova.network.neutron [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Refreshing network info cache for port 3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1128.123933] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.142s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.123933] env[62235]: DEBUG nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1128.124976] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.736s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.125217] env[62235]: DEBUG nova.objects.instance [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lazy-loading 'resources' on Instance uuid a993bed1-e346-47e7-9f7f-abdfb6685196 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.619566] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.628471] env[62235]: DEBUG nova.compute.utils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1128.633282] env[62235]: DEBUG nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1128.633467] env[62235]: DEBUG nova.network.neutron [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1128.680410] env[62235]: DEBUG nova.policy [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91b451aa2fe74f6eabee71b0f4f90f0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cd98f2307d24723b3aeb481b535ace9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1128.713643] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e8506e-61dd-4428-9451-edd5223f2aa7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.721106] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d737d13e-f079-4647-9e3b-a59db049eb41 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.761949] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b5648a-7a91-43f5-8682-383a8355cc1f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.772147] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9daae9-eb40-4951-95ff-dfbaaff712a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.787777] env[62235]: DEBUG nova.compute.provider_tree [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.836786] env[62235]: DEBUG nova.network.neutron [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updated VIF entry in instance network info cache for port 3b639a75-6787-4039-95d9-4197fe7bff26. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.837219] env[62235]: DEBUG nova.network.neutron [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap3b639a75-67", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.034548] env[62235]: DEBUG nova.network.neutron [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Successfully created port: b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1129.133880] env[62235]: DEBUG nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1129.293692] env[62235]: DEBUG nova.scheduler.client.report [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1129.343126] env[62235]: DEBUG oslo_concurrency.lockutils [req-4e2bea3c-85d1-4781-ae12-458d496aab98 req-532f6b59-7c43-4add-84d9-6badb9e62199 service nova] Releasing lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.799504] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.674s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.801814] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.623s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.802063] env[62235]: DEBUG nova.objects.instance [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lazy-loading 'resources' on Instance uuid 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.821417] env[62235]: INFO nova.scheduler.client.report [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Deleted allocations for instance a993bed1-e346-47e7-9f7f-abdfb6685196 [ 1130.143407] env[62235]: DEBUG nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1130.167839] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1130.168118] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1130.168287] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1130.168475] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1130.168629] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1130.168777] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1130.168986] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1130.169164] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1130.169371] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1130.169546] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1130.169725] env[62235]: DEBUG nova.virt.hardware [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1130.170595] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6ea46b-fcd5-43f8-8bf2-f150ba101100 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.179043] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af8f318-2e73-443a-bfbf-040ef1e2a76d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.307971] env[62235]: DEBUG nova.objects.instance [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lazy-loading 'numa_topology' on Instance uuid 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.328258] env[62235]: DEBUG oslo_concurrency.lockutils [None req-64519039-e982-4d75-98de-92360acf1081 tempest-AttachVolumeNegativeTest-1660221414 tempest-AttachVolumeNegativeTest-1660221414-project-member] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.314s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.329129] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.593s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.329326] env[62235]: INFO nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] During sync_power_state the instance has a pending task (deleting). Skip. [ 1130.329552] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "a993bed1-e346-47e7-9f7f-abdfb6685196" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.619583] env[62235]: DEBUG nova.network.neutron [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Successfully updated port: b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1130.625537] env[62235]: DEBUG nova.compute.manager [req-0226f5ae-cb4d-47d9-941e-7a50ef2e05dd req-38b85a49-b517-4fc0-8a80-82aff7b350f2 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Received event network-vif-plugged-b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.625537] env[62235]: DEBUG oslo_concurrency.lockutils [req-0226f5ae-cb4d-47d9-941e-7a50ef2e05dd req-38b85a49-b517-4fc0-8a80-82aff7b350f2 service nova] Acquiring lock "a293951c-183d-4bfe-b41e-2b3408f2eee3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.625537] env[62235]: DEBUG oslo_concurrency.lockutils [req-0226f5ae-cb4d-47d9-941e-7a50ef2e05dd req-38b85a49-b517-4fc0-8a80-82aff7b350f2 service nova] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.625537] env[62235]: DEBUG oslo_concurrency.lockutils [req-0226f5ae-cb4d-47d9-941e-7a50ef2e05dd req-38b85a49-b517-4fc0-8a80-82aff7b350f2 service nova] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.625537] env[62235]: DEBUG nova.compute.manager [req-0226f5ae-cb4d-47d9-941e-7a50ef2e05dd req-38b85a49-b517-4fc0-8a80-82aff7b350f2 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] No waiting events found dispatching network-vif-plugged-b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1130.625537] env[62235]: WARNING nova.compute.manager [req-0226f5ae-cb4d-47d9-941e-7a50ef2e05dd req-38b85a49-b517-4fc0-8a80-82aff7b350f2 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Received unexpected event network-vif-plugged-b29af1b5-b582-4e06-945e-cadfe1b86830 for instance with vm_state building and task_state spawning. [ 1130.809399] env[62235]: DEBUG nova.objects.base [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Object Instance<5e52ce21-04a1-41db-b62e-2932bf91538e> lazy-loaded attributes: resources,numa_topology {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1130.872022] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3a716a-322f-4909-bacb-11f4b7919b94 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.879105] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2ea3c4-6cb6-4968-9dca-0f79f78f0f5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.910447] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f09021-c83f-4bde-9c15-6ba4165d9054 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.917393] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d8e147-c678-49b2-8829-8b4387169740 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.930739] env[62235]: DEBUG nova.compute.provider_tree [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.123357] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.123523] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.123808] env[62235]: DEBUG nova.network.neutron [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1131.434482] env[62235]: DEBUG nova.scheduler.client.report [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.653745] env[62235]: DEBUG nova.network.neutron [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1131.773167] env[62235]: DEBUG nova.network.neutron [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updating instance_info_cache with network_info: [{"id": "b29af1b5-b582-4e06-945e-cadfe1b86830", "address": "fa:16:3e:6c:34:14", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb29af1b5-b5", "ovs_interfaceid": "b29af1b5-b582-4e06-945e-cadfe1b86830", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.938486] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.136s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.275706] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.276089] env[62235]: DEBUG nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Instance network_info: |[{"id": "b29af1b5-b582-4e06-945e-cadfe1b86830", "address": "fa:16:3e:6c:34:14", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb29af1b5-b5", "ovs_interfaceid": "b29af1b5-b582-4e06-945e-cadfe1b86830", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1132.276639] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:34:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b29af1b5-b582-4e06-945e-cadfe1b86830', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1132.284209] env[62235]: DEBUG oslo.service.loopingcall [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.284445] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1132.284686] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d46420d-bbb0-4cb7-a53b-3b936791e92b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.305529] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1132.305529] env[62235]: value = "task-1272264" [ 1132.305529] env[62235]: _type = "Task" [ 1132.305529] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.312880] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272264, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.447280] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cdbfb98f-6ab8-4444-abfb-b44ad6a6e3d4 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.266s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.447853] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.712s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.447853] env[62235]: INFO nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] During sync_power_state the instance has a pending task (shelving_image_uploading). Skip. [ 1132.448106] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.448576] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.829s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.448797] env[62235]: INFO nova.compute.manager [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Unshelving [ 1132.650508] env[62235]: DEBUG nova.compute.manager [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Received event network-changed-b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1132.650729] env[62235]: DEBUG nova.compute.manager [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Refreshing instance network info cache due to event network-changed-b29af1b5-b582-4e06-945e-cadfe1b86830. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1132.650949] env[62235]: DEBUG oslo_concurrency.lockutils [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] Acquiring lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.651112] env[62235]: DEBUG oslo_concurrency.lockutils [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] Acquired lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.651279] env[62235]: DEBUG nova.network.neutron [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Refreshing network info cache for port b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1132.815014] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272264, 'name': CreateVM_Task, 'duration_secs': 0.293229} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.815207] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1132.815899] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.816085] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.816445] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1132.816700] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4de41282-db1c-4ab0-9e95-31222a9cad9e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.820863] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1132.820863] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b12c84-12ca-0e44-e0c3-db421ad568a4" [ 1132.820863] env[62235]: _type = "Task" [ 1132.820863] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.829444] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b12c84-12ca-0e44-e0c3-db421ad568a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.330611] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b12c84-12ca-0e44-e0c3-db421ad568a4, 'name': SearchDatastore_Task, 'duration_secs': 0.009828} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.332776] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.333029] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1133.333273] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.333420] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.333604] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1133.333864] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-face1c43-a45b-43b2-b70d-756aaf803915 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.341382] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1133.341561] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1133.342241] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-152157a4-dcd8-4bdd-9586-09edf92086e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.346791] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1133.346791] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523e430b-1341-1437-8734-6aace30979a7" [ 1133.346791] env[62235]: _type = "Task" [ 1133.346791] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.353972] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523e430b-1341-1437-8734-6aace30979a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.396915] env[62235]: DEBUG nova.network.neutron [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updated VIF entry in instance network info cache for port b29af1b5-b582-4e06-945e-cadfe1b86830. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1133.397282] env[62235]: DEBUG nova.network.neutron [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updating instance_info_cache with network_info: [{"id": "b29af1b5-b582-4e06-945e-cadfe1b86830", "address": "fa:16:3e:6c:34:14", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb29af1b5-b5", "ovs_interfaceid": "b29af1b5-b582-4e06-945e-cadfe1b86830", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.473032] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.473325] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.473525] env[62235]: DEBUG nova.objects.instance [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lazy-loading 'pci_requests' on Instance uuid 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.858309] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]523e430b-1341-1437-8734-6aace30979a7, 'name': SearchDatastore_Task, 'duration_secs': 0.007924} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.859103] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d8c4be9-0ea3-49a0-8420-ff8767d2a5cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.864330] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1133.864330] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5282fe4d-8773-6bf9-b73a-fc89eb3bfdfc" [ 1133.864330] env[62235]: _type = "Task" [ 1133.864330] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.875684] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5282fe4d-8773-6bf9-b73a-fc89eb3bfdfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.899664] env[62235]: DEBUG oslo_concurrency.lockutils [req-b959e23f-e99f-4fd6-8ed1-fc615e0def3c req-41d19d73-494f-48b5-9747-d2adbe8e0cbc service nova] Releasing lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.977203] env[62235]: DEBUG nova.objects.instance [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lazy-loading 'numa_topology' on Instance uuid 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.376302] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5282fe4d-8773-6bf9-b73a-fc89eb3bfdfc, 'name': SearchDatastore_Task, 'duration_secs': 0.009394} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.376644] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.376922] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1134.377222] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1abb8f5a-c3df-4a06-92a2-b18179edfe10 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.384550] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1134.384550] env[62235]: value = "task-1272266" [ 1134.384550] env[62235]: _type = "Task" [ 1134.384550] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.392998] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272266, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.479828] env[62235]: INFO nova.compute.claims [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1134.894705] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272266, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.41949} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.894999] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1134.895235] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1134.895484] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15572651-559b-4ea3-87f6-652450713a80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.902011] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1134.902011] env[62235]: value = "task-1272267" [ 1134.902011] env[62235]: _type = "Task" [ 1134.902011] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.909695] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272267, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.412973] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272267, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071877} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.413265] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1135.413999] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a3fbd1-5b1c-4431-b91e-df92358d37ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.444643] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.444971] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6eb39c8a-ee5d-469c-9c15-64b24b2ecbff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.472900] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1135.472900] env[62235]: value = "task-1272268" [ 1135.472900] env[62235]: _type = "Task" [ 1135.472900] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.481785] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272268, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.554014] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bd2b57-4ded-45fe-bad2-2dddc3d1e4a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.561526] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe1d00f-347c-4287-b98d-8efe8f879c05 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.594144] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb4ea99-2853-4032-bf09-3aff77b41000 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.602384] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cd9ca9-c5cf-40e7-b206-ebab61e9f8a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.617199] env[62235]: DEBUG nova.compute.provider_tree [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.982469] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272268, 'name': ReconfigVM_Task, 'duration_secs': 0.268845} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.982778] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfigured VM instance instance-0000006a to attach disk [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.983732] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3604394-28c1-45ae-be03-ad5596b72151 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.990982] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1135.990982] env[62235]: value = "task-1272269" [ 1135.990982] env[62235]: _type = "Task" [ 1135.990982] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.000015] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272269, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.120027] env[62235]: DEBUG nova.scheduler.client.report [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.500676] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272269, 'name': Rename_Task, 'duration_secs': 0.136766} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.500964] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1136.501225] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf1ce28c-0684-4ac7-985e-4bb727ae2543 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.507913] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1136.507913] env[62235]: value = "task-1272271" [ 1136.507913] env[62235]: _type = "Task" [ 1136.507913] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.515329] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272271, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.625042] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.151s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.652469] env[62235]: INFO nova.network.neutron [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating port 3b639a75-6787-4039-95d9-4197fe7bff26 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1137.017587] env[62235]: DEBUG oslo_vmware.api [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272271, 'name': PowerOnVM_Task, 'duration_secs': 0.445392} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.017977] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1137.018240] env[62235]: INFO nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Took 6.87 seconds to spawn the instance on the hypervisor. [ 1137.018434] env[62235]: DEBUG nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1137.019194] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032416c9-383f-407a-bc52-7070908db5d0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.535030] env[62235]: INFO nova.compute.manager [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Took 12.33 seconds to build instance. [ 1138.039658] env[62235]: DEBUG oslo_concurrency.lockutils [None req-8280d6df-5ab1-4379-ab56-4489df04fdf3 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.843s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.066302] env[62235]: DEBUG nova.compute.manager [req-6f6ffdda-b17d-4f49-afdb-6bd02df1e4bc req-51c21ad4-ac41-4374-809a-1d5fb6103577 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received event network-vif-plugged-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.066498] env[62235]: DEBUG oslo_concurrency.lockutils [req-6f6ffdda-b17d-4f49-afdb-6bd02df1e4bc req-51c21ad4-ac41-4374-809a-1d5fb6103577 service nova] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.066715] env[62235]: DEBUG oslo_concurrency.lockutils [req-6f6ffdda-b17d-4f49-afdb-6bd02df1e4bc req-51c21ad4-ac41-4374-809a-1d5fb6103577 service nova] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.066898] env[62235]: DEBUG oslo_concurrency.lockutils [req-6f6ffdda-b17d-4f49-afdb-6bd02df1e4bc req-51c21ad4-ac41-4374-809a-1d5fb6103577 service nova] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.067614] env[62235]: DEBUG nova.compute.manager [req-6f6ffdda-b17d-4f49-afdb-6bd02df1e4bc req-51c21ad4-ac41-4374-809a-1d5fb6103577 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] No waiting events found dispatching network-vif-plugged-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1138.067840] env[62235]: WARNING nova.compute.manager [req-6f6ffdda-b17d-4f49-afdb-6bd02df1e4bc req-51c21ad4-ac41-4374-809a-1d5fb6103577 service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received unexpected event network-vif-plugged-3b639a75-6787-4039-95d9-4197fe7bff26 for instance with vm_state shelved_offloaded and task_state spawning. [ 1138.156975] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.159113] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.159348] env[62235]: DEBUG nova.network.neutron [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.410473] env[62235]: DEBUG nova.compute.manager [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Received event network-changed-b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.410648] env[62235]: DEBUG nova.compute.manager [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Refreshing instance network info cache due to event network-changed-b29af1b5-b582-4e06-945e-cadfe1b86830. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1138.410855] env[62235]: DEBUG oslo_concurrency.lockutils [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] Acquiring lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.411011] env[62235]: DEBUG oslo_concurrency.lockutils [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] Acquired lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.411195] env[62235]: DEBUG nova.network.neutron [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Refreshing network info cache for port b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1138.885085] env[62235]: DEBUG nova.network.neutron [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b639a75-67", "ovs_interfaceid": "3b639a75-6787-4039-95d9-4197fe7bff26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.114793] env[62235]: DEBUG nova.network.neutron [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updated VIF entry in instance network info cache for port b29af1b5-b582-4e06-945e-cadfe1b86830. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1139.115164] env[62235]: DEBUG nova.network.neutron [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updating instance_info_cache with network_info: [{"id": "b29af1b5-b582-4e06-945e-cadfe1b86830", "address": "fa:16:3e:6c:34:14", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.223", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb29af1b5-b5", "ovs_interfaceid": "b29af1b5-b582-4e06-945e-cadfe1b86830", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.388382] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.415691] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b69777c1a0e7f5166c9bf01cc427c1d7',container_format='bare',created_at=2024-10-12T06:28:11Z,direct_url=,disk_format='vmdk',id=23fc41d5-108d-47df-b47c-d7f47435d593,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-494343846-shelved',owner='ab1aacbb5a8746769c04f2df5f99d01a',properties=ImageMetaProps,protected=,size=31596032,status='active',tags=,updated_at=2024-10-12T06:28:27Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.415691] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.415930] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.416256] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.416980] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.417229] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.417513] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.417728] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.417961] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.418165] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.418371] env[62235]: DEBUG nova.virt.hardware [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.419382] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ca9c6e-2cfb-43de-b0f3-00878e463dc6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.428802] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d8d9af-35de-49bc-87e6-7cbc35dd0f43 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.445937] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:3c:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73eeba7c-29e1-4fdf-82b3-d62e63e86051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b639a75-6787-4039-95d9-4197fe7bff26', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.454593] env[62235]: DEBUG oslo.service.loopingcall [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.454884] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1139.455152] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-450fed6b-b8cc-4921-839e-13c439df631f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.476682] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1139.476682] env[62235]: value = "task-1272272" [ 1139.476682] env[62235]: _type = "Task" [ 1139.476682] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.484777] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272272, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.617910] env[62235]: DEBUG oslo_concurrency.lockutils [req-1e4a0cdc-d40d-435d-a70e-b1ea5a192ef4 req-5ddc5f2a-8508-42c7-ae4d-725530566bb8 service nova] Releasing lock "refresh_cache-a293951c-183d-4bfe-b41e-2b3408f2eee3" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.986847] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272272, 'name': CreateVM_Task, 'duration_secs': 0.292506} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.987462] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1139.988264] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.989326] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.989326] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1139.989326] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07d2968e-4da9-47d8-ab67-cc6dc3b62b03 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.994399] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1139.994399] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ebaa88-42d4-5a6c-2344-f01aa5ffa32e" [ 1139.994399] env[62235]: _type = "Task" [ 1139.994399] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.003252] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ebaa88-42d4-5a6c-2344-f01aa5ffa32e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.096044] env[62235]: DEBUG nova.compute.manager [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received event network-changed-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.096269] env[62235]: DEBUG nova.compute.manager [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Refreshing instance network info cache due to event network-changed-3b639a75-6787-4039-95d9-4197fe7bff26. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1140.096489] env[62235]: DEBUG oslo_concurrency.lockutils [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] Acquiring lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.096636] env[62235]: DEBUG oslo_concurrency.lockutils [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] Acquired lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.096799] env[62235]: DEBUG nova.network.neutron [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Refreshing network info cache for port 3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1140.504655] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.505355] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Processing image 23fc41d5-108d-47df-b47c-d7f47435d593 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1140.505947] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593/23fc41d5-108d-47df-b47c-d7f47435d593.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.506314] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593/23fc41d5-108d-47df-b47c-d7f47435d593.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.506744] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1140.507379] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7fec557-abd3-46e9-8609-49f069a0ea02 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.515857] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1140.517079] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1140.517079] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a03fd7ad-24df-4fea-bfbf-ade70fe929b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.522380] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1140.522380] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52de5bb3-b05f-9592-9442-adc5d2d9e798" [ 1140.522380] env[62235]: _type = "Task" [ 1140.522380] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.529995] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52de5bb3-b05f-9592-9442-adc5d2d9e798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.821281] env[62235]: DEBUG nova.network.neutron [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updated VIF entry in instance network info cache for port 3b639a75-6787-4039-95d9-4197fe7bff26. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1140.821692] env[62235]: DEBUG nova.network.neutron [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b639a75-67", "ovs_interfaceid": "3b639a75-6787-4039-95d9-4197fe7bff26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.031868] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Preparing fetch location {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1141.032508] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Fetch image to [datastore1] OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09/OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09.vmdk {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1141.032907] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Downloading stream optimized image 23fc41d5-108d-47df-b47c-d7f47435d593 to [datastore1] OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09/OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09.vmdk on the data store datastore1 as vApp {{(pid=62235) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1141.033288] env[62235]: DEBUG nova.virt.vmwareapi.images [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Downloading image file data 23fc41d5-108d-47df-b47c-d7f47435d593 to the ESX as VM named 'OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09' {{(pid=62235) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1141.117259] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1141.117259] env[62235]: value = "resgroup-9" [ 1141.117259] env[62235]: _type = "ResourcePool" [ 1141.117259] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1141.117485] env[62235]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5671f60d-ad2a-461b-866e-cfd0582a3397 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.138661] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lease: (returnval){ [ 1141.138661] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232cc76-c46e-016e-11d7-6e6f371b8c2b" [ 1141.138661] env[62235]: _type = "HttpNfcLease" [ 1141.138661] env[62235]: } obtained for vApp import into resource pool (val){ [ 1141.138661] env[62235]: value = "resgroup-9" [ 1141.138661] env[62235]: _type = "ResourcePool" [ 1141.138661] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1141.138986] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the lease: (returnval){ [ 1141.138986] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232cc76-c46e-016e-11d7-6e6f371b8c2b" [ 1141.138986] env[62235]: _type = "HttpNfcLease" [ 1141.138986] env[62235]: } to be ready. {{(pid=62235) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1141.145151] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1141.145151] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232cc76-c46e-016e-11d7-6e6f371b8c2b" [ 1141.145151] env[62235]: _type = "HttpNfcLease" [ 1141.145151] env[62235]: } is initializing. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1141.324418] env[62235]: DEBUG oslo_concurrency.lockutils [req-797d6bb1-c976-45d0-9c0b-27706fb22da8 req-915f50e7-4e0e-40fc-ac5c-0bbb8d3d8d8f service nova] Releasing lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.646782] env[62235]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1141.646782] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232cc76-c46e-016e-11d7-6e6f371b8c2b" [ 1141.646782] env[62235]: _type = "HttpNfcLease" [ 1141.646782] env[62235]: } is ready. {{(pid=62235) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1141.647102] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1141.647102] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5232cc76-c46e-016e-11d7-6e6f371b8c2b" [ 1141.647102] env[62235]: _type = "HttpNfcLease" [ 1141.647102] env[62235]: }. {{(pid=62235) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1141.647816] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb07584-f63e-46dd-ad3e-d6de08bf1606 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.655101] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522084a5-3f36-50f7-b9bf-83b65ad13a3f/disk-0.vmdk from lease info. {{(pid=62235) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1141.655278] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating HTTP connection to write to file with size = 31596032 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522084a5-3f36-50f7-b9bf-83b65ad13a3f/disk-0.vmdk. {{(pid=62235) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1141.719796] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5c7072ce-2610-4572-b8b8-894d1d16337c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.764456] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Completed reading data from the image iterator. {{(pid=62235) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1142.764807] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522084a5-3f36-50f7-b9bf-83b65ad13a3f/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1142.765798] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a4b84e-8b46-4a54-aa90-4907858ae013 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.774735] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522084a5-3f36-50f7-b9bf-83b65ad13a3f/disk-0.vmdk is in state: ready. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1142.774946] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522084a5-3f36-50f7-b9bf-83b65ad13a3f/disk-0.vmdk. {{(pid=62235) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1142.775224] env[62235]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9f237c7a-4a8f-4856-8a5f-fbeecf8674fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.484136] env[62235]: DEBUG oslo_vmware.rw_handles [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522084a5-3f36-50f7-b9bf-83b65ad13a3f/disk-0.vmdk. {{(pid=62235) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1143.484381] env[62235]: INFO nova.virt.vmwareapi.images [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Downloaded image file data 23fc41d5-108d-47df-b47c-d7f47435d593 [ 1143.485224] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683390e8-acc1-44c0-9370-ebc8df81034b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.499893] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c78df5d9-29da-4a20-963c-c1dc4c4434ec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.697064] env[62235]: INFO nova.virt.vmwareapi.images [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] The imported VM was unregistered [ 1143.699450] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Caching image {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1143.699689] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Creating directory with path [datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1143.699971] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c38ca528-7e68-4552-ba8b-5e0bc6b7ed58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.721372] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Created directory with path [datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593 {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1143.721581] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09/OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09.vmdk to [datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593/23fc41d5-108d-47df-b47c-d7f47435d593.vmdk. {{(pid=62235) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1143.721846] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-289d5148-689d-448e-a092-cb3b61b3faea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.728770] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1143.728770] env[62235]: value = "task-1272275" [ 1143.728770] env[62235]: _type = "Task" [ 1143.728770] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.736831] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272275, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.238969] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272275, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.739375] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272275, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.241086] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272275, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.740794] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272275, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.242056] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272275, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.337289} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.242433] env[62235]: INFO nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09/OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09.vmdk to [datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593/23fc41d5-108d-47df-b47c-d7f47435d593.vmdk. [ 1146.242547] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Cleaning up location [datastore1] OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1146.242721] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_dad04de4-7076-4cee-b1d6-caa955eb6e09 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1146.242972] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a0215f15-59ab-4673-8910-dd7472ae4694 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.250847] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1146.250847] env[62235]: value = "task-1272276" [ 1146.250847] env[62235]: _type = "Task" [ 1146.250847] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.257926] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272276, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.761023] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272276, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035812} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.761364] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1146.761488] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593/23fc41d5-108d-47df-b47c-d7f47435d593.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.761746] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593/23fc41d5-108d-47df-b47c-d7f47435d593.vmdk to [datastore1] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1146.761999] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1529d9c-67c5-4a99-a602-83dc999e9f74 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.769424] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1146.769424] env[62235]: value = "task-1272277" [ 1146.769424] env[62235]: _type = "Task" [ 1146.769424] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.777069] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.279193] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272277, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.780532] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272277, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.281638] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272277, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.782528] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272277, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.282856] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272277, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.128597} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.283221] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/23fc41d5-108d-47df-b47c-d7f47435d593/23fc41d5-108d-47df-b47c-d7f47435d593.vmdk to [datastore1] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1149.283964] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d22df43-dfc4-4a0c-932c-178b9fe3cb78 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.305992] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.306246] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-737cbb6d-7beb-4fe0-838e-c864091cf540 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.326084] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1149.326084] env[62235]: value = "task-1272278" [ 1149.326084] env[62235]: _type = "Task" [ 1149.326084] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.335330] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272278, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.836417] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272278, 'name': ReconfigVM_Task, 'duration_secs': 0.297018} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.836695] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 5e52ce21-04a1-41db-b62e-2932bf91538e/5e52ce21-04a1-41db-b62e-2932bf91538e.vmdk or device None with type streamOptimized {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1149.837319] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2a6a86f-40fd-49fe-a49d-826c805b3975 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.843177] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1149.843177] env[62235]: value = "task-1272279" [ 1149.843177] env[62235]: _type = "Task" [ 1149.843177] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.850601] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272279, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.887171] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "f7ef033e-ab43-4b89-b331-cb234a373991" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.887453] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.887645] env[62235]: DEBUG nova.compute.manager [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.888421] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5a4414-c852-4460-8056-30ec0c433ba7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.894364] env[62235]: DEBUG nova.compute.manager [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62235) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1149.894955] env[62235]: DEBUG nova.objects.instance [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'flavor' on Instance uuid f7ef033e-ab43-4b89-b331-cb234a373991 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1150.353414] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272279, 'name': Rename_Task, 'duration_secs': 0.135852} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.353779] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1150.353874] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d8a79b5e-6f30-4a2f-9aab-8df69d9a53bd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.360794] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1150.360794] env[62235]: value = "task-1272280" [ 1150.360794] env[62235]: _type = "Task" [ 1150.360794] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.367779] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272280, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.400524] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1150.400772] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9871a1d-8c6a-4e6f-af1d-0c7b621f367e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.406650] env[62235]: DEBUG oslo_vmware.api [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1150.406650] env[62235]: value = "task-1272281" [ 1150.406650] env[62235]: _type = "Task" [ 1150.406650] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.414478] env[62235]: DEBUG oslo_vmware.api [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272281, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.870149] env[62235]: DEBUG oslo_vmware.api [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272280, 'name': PowerOnVM_Task, 'duration_secs': 0.451004} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.870411] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1150.916115] env[62235]: DEBUG oslo_vmware.api [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272281, 'name': PowerOffVM_Task, 'duration_secs': 0.193527} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.916416] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1150.916610] env[62235]: DEBUG nova.compute.manager [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1150.917407] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f935f14-20b7-4322-9e60-4523b219a066 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.965560] env[62235]: DEBUG nova.compute.manager [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1150.966464] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def6a2ab-2888-4474-b6e3-3738eaea03f3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.428981] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4a740a2f-d3b3-449c-ad6e-da2bd7d0f3f4 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.541s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.483928] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4c506ee-27a2-4da6-8fb5-c93ff03287eb tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.035s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.287285] env[62235]: DEBUG nova.objects.instance [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'flavor' on Instance uuid f7ef033e-ab43-4b89-b331-cb234a373991 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.544867] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de8107b-75b7-4947-9564-8e5d50aaf394 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.551870] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fda4fa1d-c97f-4a30-8a92-57e5f88a17c2 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Suspending the VM {{(pid=62235) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1152.552133] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e10a042a-841d-4ed4-a0ec-ee5a56b8d655 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.559301] env[62235]: DEBUG oslo_vmware.api [None req-fda4fa1d-c97f-4a30-8a92-57e5f88a17c2 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1152.559301] env[62235]: value = "task-1272282" [ 1152.559301] env[62235]: _type = "Task" [ 1152.559301] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.566961] env[62235]: DEBUG oslo_vmware.api [None req-fda4fa1d-c97f-4a30-8a92-57e5f88a17c2 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272282, 'name': SuspendVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.792586] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.792897] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.793221] env[62235]: DEBUG nova.network.neutron [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1152.793548] env[62235]: DEBUG nova.objects.instance [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'info_cache' on Instance uuid f7ef033e-ab43-4b89-b331-cb234a373991 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.070172] env[62235]: DEBUG oslo_vmware.api [None req-fda4fa1d-c97f-4a30-8a92-57e5f88a17c2 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272282, 'name': SuspendVM_Task} progress is 100%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.296759] env[62235]: DEBUG nova.objects.base [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1153.570339] env[62235]: DEBUG oslo_vmware.api [None req-fda4fa1d-c97f-4a30-8a92-57e5f88a17c2 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272282, 'name': SuspendVM_Task, 'duration_secs': 0.517292} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.570700] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-fda4fa1d-c97f-4a30-8a92-57e5f88a17c2 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Suspended the VM {{(pid=62235) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1153.570814] env[62235]: DEBUG nova.compute.manager [None req-fda4fa1d-c97f-4a30-8a92-57e5f88a17c2 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1153.571550] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a1ef28-7e2f-4c4a-929b-8639375194d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.996023] env[62235]: DEBUG nova.network.neutron [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updating instance_info_cache with network_info: [{"id": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "address": "fa:16:3e:c0:27:54", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape017afa0-7a", "ovs_interfaceid": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.498443] env[62235]: DEBUG oslo_concurrency.lockutils [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.929892] env[62235]: INFO nova.compute.manager [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Resuming [ 1154.930604] env[62235]: DEBUG nova.objects.instance [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lazy-loading 'flavor' on Instance uuid 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.001565] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1155.001852] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-142bf10e-35f0-4f7f-bcfd-14cafbaf928c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.009749] env[62235]: DEBUG oslo_vmware.api [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1155.009749] env[62235]: value = "task-1272283" [ 1155.009749] env[62235]: _type = "Task" [ 1155.009749] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.017236] env[62235]: DEBUG oslo_vmware.api [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.520330] env[62235]: DEBUG oslo_vmware.api [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272283, 'name': PowerOnVM_Task, 'duration_secs': 0.336235} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.520554] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1155.520743] env[62235]: DEBUG nova.compute.manager [None req-51434fd8-4bf5-4443-9e14-18cafe9485f1 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1155.521482] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49246238-6378-4fba-96b5-3d66370c59b3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.938955] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.939338] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquired lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.939378] env[62235]: DEBUG nova.network.neutron [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1156.662568] env[62235]: DEBUG nova.network.neutron [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [{"id": "3b639a75-6787-4039-95d9-4197fe7bff26", "address": "fa:16:3e:0e:3c:fe", "network": {"id": "a7b0214e-2011-427e-887a-1e5c62f19d7f", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1175419132-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab1aacbb5a8746769c04f2df5f99d01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b639a75-67", "ovs_interfaceid": "3b639a75-6787-4039-95d9-4197fe7bff26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.165388] env[62235]: DEBUG oslo_concurrency.lockutils [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Releasing lock "refresh_cache-5e52ce21-04a1-41db-b62e-2932bf91538e" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.167738] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e9edc0-8ebe-4e72-ac44-fdb2e5f50eca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.174701] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Resuming the VM {{(pid=62235) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1157.175012] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bfdbe1ad-dc90-43a8-b9dc-75eb80b541a2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.181164] env[62235]: DEBUG oslo_vmware.api [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1157.181164] env[62235]: value = "task-1272284" [ 1157.181164] env[62235]: _type = "Task" [ 1157.181164] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.189074] env[62235]: DEBUG oslo_vmware.api [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.549439] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d46017-e7f5-427e-869f-907a36dcbe91 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.556257] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3b33fc82-0f87-4687-9ccf-a799b9ffbd8e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Suspending the VM {{(pid=62235) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1157.556476] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-58932073-3762-4feb-b3b5-8f7141523e09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.563028] env[62235]: DEBUG oslo_vmware.api [None req-3b33fc82-0f87-4687-9ccf-a799b9ffbd8e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1157.563028] env[62235]: value = "task-1272285" [ 1157.563028] env[62235]: _type = "Task" [ 1157.563028] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.571208] env[62235]: DEBUG oslo_vmware.api [None req-3b33fc82-0f87-4687-9ccf-a799b9ffbd8e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272285, 'name': SuspendVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.691092] env[62235]: DEBUG oslo_vmware.api [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272284, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.074945] env[62235]: DEBUG oslo_vmware.api [None req-3b33fc82-0f87-4687-9ccf-a799b9ffbd8e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272285, 'name': SuspendVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.191986] env[62235]: DEBUG oslo_vmware.api [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272284, 'name': PowerOnVM_Task, 'duration_secs': 0.860938} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.192414] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Resumed the VM {{(pid=62235) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1158.192596] env[62235]: DEBUG nova.compute.manager [None req-75646d05-81b8-4cf1-af2d-e75a85d5227e tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.193445] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b112b9b-1a9f-4a7f-bd25-389707fb22be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.574495] env[62235]: DEBUG oslo_vmware.api [None req-3b33fc82-0f87-4687-9ccf-a799b9ffbd8e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272285, 'name': SuspendVM_Task, 'duration_secs': 0.865593} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.574855] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-3b33fc82-0f87-4687-9ccf-a799b9ffbd8e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Suspended the VM {{(pid=62235) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1158.575103] env[62235]: DEBUG nova.compute.manager [None req-3b33fc82-0f87-4687-9ccf-a799b9ffbd8e tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.575973] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7252404e-fef1-4b75-8391-0264e21763ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.945914] env[62235]: INFO nova.compute.manager [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Resuming [ 1159.946621] env[62235]: DEBUG nova.objects.instance [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'flavor' on Instance uuid f7ef033e-ab43-4b89-b331-cb234a373991 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.955289] env[62235]: DEBUG oslo_concurrency.lockutils [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.955660] env[62235]: DEBUG oslo_concurrency.lockutils [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquired lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.955660] env[62235]: DEBUG nova.network.neutron [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1161.663292] env[62235]: DEBUG nova.network.neutron [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updating instance_info_cache with network_info: [{"id": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "address": "fa:16:3e:c0:27:54", "network": {"id": "6b295ada-d077-411a-8992-49f74adc8cf4", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1309358242-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c943a445f67f4021bad11723e0836537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37fb1918-d178-4e12-93e6-316381e78be4", "external-id": "nsx-vlan-transportzone-763", "segmentation_id": 763, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape017afa0-7a", "ovs_interfaceid": "e017afa0-7a52-4f06-8ec1-5e7e24fb337a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.166355] env[62235]: DEBUG oslo_concurrency.lockutils [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Releasing lock "refresh_cache-f7ef033e-ab43-4b89-b331-cb234a373991" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.167676] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a372c47c-ad91-499a-a228-52d76c92b4a7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.175860] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Resuming the VM {{(pid=62235) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1162.176165] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5f26ce4-d787-4537-a3b8-532943062684 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.183116] env[62235]: DEBUG oslo_vmware.api [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1162.183116] env[62235]: value = "task-1272286" [ 1162.183116] env[62235]: _type = "Task" [ 1162.183116] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.192755] env[62235]: DEBUG oslo_vmware.api [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272286, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.699269] env[62235]: DEBUG oslo_vmware.api [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272286, 'name': PowerOnVM_Task, 'duration_secs': 0.481982} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.699493] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Resumed the VM {{(pid=62235) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1162.699688] env[62235]: DEBUG nova.compute.manager [None req-36c31104-e9a2-48a7-a013-d60532b954c9 tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1162.700574] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa0f768-7118-454f-a69b-cefe74349743 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.255383] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.255752] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.255853] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.256056] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.256236] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.258336] env[62235]: INFO nova.compute.manager [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Terminating instance [ 1163.260052] env[62235]: DEBUG nova.compute.manager [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1163.260286] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1163.261340] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e108a05f-893e-467c-a91d-70fa97aa04da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.268865] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1163.269112] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98d84d38-557a-4e14-a0be-637cf12650c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.276265] env[62235]: DEBUG oslo_vmware.api [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1163.276265] env[62235]: value = "task-1272287" [ 1163.276265] env[62235]: _type = "Task" [ 1163.276265] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.283294] env[62235]: DEBUG oslo_vmware.api [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.613075] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "f7ef033e-ab43-4b89-b331-cb234a373991" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.613287] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.613441] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "f7ef033e-ab43-4b89-b331-cb234a373991-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.613635] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.613812] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.616044] env[62235]: INFO nova.compute.manager [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Terminating instance [ 1163.617949] env[62235]: DEBUG nova.compute.manager [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1163.618163] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1163.618975] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a6f25c-dd70-4e8e-8f04-b2142a7907c6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.626344] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1163.626565] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b545cbf1-e88a-48b9-9c16-4fef020ff84f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.631992] env[62235]: DEBUG oslo_vmware.api [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1163.631992] env[62235]: value = "task-1272288" [ 1163.631992] env[62235]: _type = "Task" [ 1163.631992] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.638916] env[62235]: DEBUG oslo_vmware.api [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.788408] env[62235]: DEBUG oslo_vmware.api [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272287, 'name': PowerOffVM_Task, 'duration_secs': 0.196881} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.788842] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1163.789155] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1163.789535] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3df3ecfe-f6c2-415c-8e63-26201e80e515 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.847855] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1163.848079] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1163.848275] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleting the datastore file [datastore1] 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.848612] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-806563e7-2f63-47f1-aaba-6a60c0b977a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.855198] env[62235]: DEBUG oslo_vmware.api [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for the task: (returnval){ [ 1163.855198] env[62235]: value = "task-1272290" [ 1163.855198] env[62235]: _type = "Task" [ 1163.855198] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.863033] env[62235]: DEBUG oslo_vmware.api [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.141922] env[62235]: DEBUG oslo_vmware.api [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272288, 'name': PowerOffVM_Task, 'duration_secs': 0.168081} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.142262] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1164.142494] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1164.142780] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7723755-c74e-4b56-8a1f-2445b7c1703c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.199962] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1164.200212] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1164.200407] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleting the datastore file [datastore2] f7ef033e-ab43-4b89-b331-cb234a373991 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1164.200675] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7e44a23-0337-4039-b6f8-6e868673e75b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.206927] env[62235]: DEBUG oslo_vmware.api [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for the task: (returnval){ [ 1164.206927] env[62235]: value = "task-1272292" [ 1164.206927] env[62235]: _type = "Task" [ 1164.206927] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.214220] env[62235]: DEBUG oslo_vmware.api [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.365060] env[62235]: DEBUG oslo_vmware.api [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Task: {'id': task-1272290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139052} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.365479] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.365541] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1164.365725] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1164.365904] env[62235]: INFO nova.compute.manager [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1164.366178] env[62235]: DEBUG oslo.service.loopingcall [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.366381] env[62235]: DEBUG nova.compute.manager [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.366475] env[62235]: DEBUG nova.network.neutron [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1164.606868] env[62235]: DEBUG nova.compute.manager [req-4b7d10db-ab48-4150-9971-d934891ab6e2 req-f82ff4d7-19c6-426e-a4db-39cfad95f57f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Received event network-vif-deleted-3b639a75-6787-4039-95d9-4197fe7bff26 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1164.607055] env[62235]: INFO nova.compute.manager [req-4b7d10db-ab48-4150-9971-d934891ab6e2 req-f82ff4d7-19c6-426e-a4db-39cfad95f57f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Neutron deleted interface 3b639a75-6787-4039-95d9-4197fe7bff26; detaching it from the instance and deleting it from the info cache [ 1164.607242] env[62235]: DEBUG nova.network.neutron [req-4b7d10db-ab48-4150-9971-d934891ab6e2 req-f82ff4d7-19c6-426e-a4db-39cfad95f57f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.716908] env[62235]: DEBUG oslo_vmware.api [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Task: {'id': task-1272292, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139996} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.717129] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.717323] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1164.717504] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1164.717677] env[62235]: INFO nova.compute.manager [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1164.717930] env[62235]: DEBUG oslo.service.loopingcall [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.718142] env[62235]: DEBUG nova.compute.manager [-] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.718240] env[62235]: DEBUG nova.network.neutron [-] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1165.087996] env[62235]: DEBUG nova.network.neutron [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.110591] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d04baafd-f454-4248-93c7-fab7eb9c61b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.120315] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec8c1f9-1893-4d36-9532-a8c27c4c3d55 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.142545] env[62235]: DEBUG nova.compute.manager [req-4b7d10db-ab48-4150-9971-d934891ab6e2 req-f82ff4d7-19c6-426e-a4db-39cfad95f57f service nova] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Detach interface failed, port_id=3b639a75-6787-4039-95d9-4197fe7bff26, reason: Instance 5e52ce21-04a1-41db-b62e-2932bf91538e could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1165.590619] env[62235]: INFO nova.compute.manager [-] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Took 1.22 seconds to deallocate network for instance. [ 1165.601883] env[62235]: DEBUG nova.network.neutron [-] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.097712] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.098027] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.098305] env[62235]: DEBUG nova.objects.instance [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lazy-loading 'resources' on Instance uuid 5e52ce21-04a1-41db-b62e-2932bf91538e {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.104064] env[62235]: INFO nova.compute.manager [-] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Took 1.39 seconds to deallocate network for instance. [ 1166.609426] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.630663] env[62235]: DEBUG nova.compute.manager [req-b51e3478-9a62-40d6-8094-81dfcbf128fa req-41839e89-faa8-4df6-82cf-f521000e9d4b service nova] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Received event network-vif-deleted-e017afa0-7a52-4f06-8ec1-5e7e24fb337a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1166.660841] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9593da2-1127-4de0-9f13-47e6824c2745 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.668779] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d97bd7-549c-42fb-9d28-60ecff98e076 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.697696] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591aa50e-2d3e-44a9-94cc-66ca2ffd6dcc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.704725] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0839eb46-88ee-4a15-a3d0-3dd5d7a90bf0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.717270] env[62235]: DEBUG nova.compute.provider_tree [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.220463] env[62235]: DEBUG nova.scheduler.client.report [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.725629] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.627s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.727842] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.119s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.728098] env[62235]: DEBUG nova.objects.instance [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lazy-loading 'resources' on Instance uuid f7ef033e-ab43-4b89-b331-cb234a373991 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.745055] env[62235]: INFO nova.scheduler.client.report [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Deleted allocations for instance 5e52ce21-04a1-41db-b62e-2932bf91538e [ 1168.252369] env[62235]: DEBUG oslo_concurrency.lockutils [None req-41d0eae3-b7a4-48a8-ab66-5fdec762cee0 tempest-ServersNegativeTestJSON-916030395 tempest-ServersNegativeTestJSON-916030395-project-member] Lock "5e52ce21-04a1-41db-b62e-2932bf91538e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.997s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.277033] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6767a947-7fb5-4798-a52b-1dc94c0e3a83 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.284520] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8972d39-3cb7-4024-8e34-39d5881e6f59 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.314216] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925ee997-d5b5-4f64-88b1-6415c7b394f8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.321393] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25526ded-90f9-4a38-90b1-f75b6bc93637 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.334264] env[62235]: DEBUG nova.compute.provider_tree [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.837390] env[62235]: DEBUG nova.scheduler.client.report [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1169.341925] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.368956] env[62235]: INFO nova.scheduler.client.report [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Deleted allocations for instance f7ef033e-ab43-4b89-b331-cb234a373991 [ 1169.878681] env[62235]: DEBUG oslo_concurrency.lockutils [None req-e4da0ef1-f79e-4a60-bacf-952d13c327ad tempest-ServerActionsTestJSON-552300181 tempest-ServerActionsTestJSON-552300181-project-member] Lock "f7ef033e-ab43-4b89-b331-cb234a373991" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.265s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.271432] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "b39057fd-7acc-437a-bb61-4b84ba4df435" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.271732] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.774392] env[62235]: DEBUG nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1174.844837] env[62235]: DEBUG oslo_concurrency.lockutils [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.845087] env[62235]: DEBUG oslo_concurrency.lockutils [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.299220] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.299524] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.300945] env[62235]: INFO nova.compute.claims [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1175.348285] env[62235]: DEBUG nova.compute.utils [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1175.850868] env[62235]: DEBUG oslo_concurrency.lockutils [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.355383] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5131ef-d06f-45ef-9862-3a3aa8057300 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.362946] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fb5b16-31ee-47e6-b8f8-3d13cbbf0886 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.394641] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ea56c1-b597-414a-8b38-0343b17bb990 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.402707] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c70dd5d-4167-47f9-b938-a046e4d1d0ea {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.419889] env[62235]: DEBUG nova.compute.provider_tree [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.922211] env[62235]: DEBUG oslo_concurrency.lockutils [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.922536] env[62235]: DEBUG oslo_concurrency.lockutils [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.922814] env[62235]: INFO nova.compute.manager [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Attaching volume 6c24557c-9453-4c92-b5bc-4255c395522e to /dev/sdb [ 1176.925160] env[62235]: DEBUG nova.scheduler.client.report [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.957022] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ff5557-dd1e-4637-abba-d32057fd58b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.964052] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f86566-7e3b-4d2b-b070-e52e9af279cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.976829] env[62235]: DEBUG nova.virt.block_device [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updating existing volume attachment record: f0aaf2c5-283f-4b2d-a068-59d45e31de5a {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1177.430626] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.130s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.430993] env[62235]: DEBUG nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1177.936147] env[62235]: DEBUG nova.compute.utils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1177.937542] env[62235]: DEBUG nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1177.937713] env[62235]: DEBUG nova.network.neutron [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1177.985138] env[62235]: DEBUG nova.policy [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ebc4275ded624a1f9bd9ed8d708aa9e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf6f966ac21f49f08bcd0f413f1d6e72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1178.220884] env[62235]: DEBUG nova.network.neutron [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Successfully created port: 7c34d6b4-c887-439a-8c23-72d06b575f7c {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1178.441037] env[62235]: DEBUG nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1179.450282] env[62235]: DEBUG nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1179.475975] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1179.476244] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1179.476408] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1179.476599] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1179.477157] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1179.477157] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1179.477157] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1179.477313] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1179.477432] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1179.477597] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1179.477769] env[62235]: DEBUG nova.virt.hardware [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1179.478661] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf318c8d-4041-4a5c-8628-a382f1eb7fa3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.486885] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4578f5-968c-45d2-ae65-89fd906df2f3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.668236] env[62235]: DEBUG nova.compute.manager [req-e3a3dc40-eae6-4d8a-a19d-95b0d9d8ecd1 req-2be7dfde-e739-4b2a-8f63-0b13e239e8cd service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Received event network-vif-plugged-7c34d6b4-c887-439a-8c23-72d06b575f7c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1179.668516] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a3dc40-eae6-4d8a-a19d-95b0d9d8ecd1 req-2be7dfde-e739-4b2a-8f63-0b13e239e8cd service nova] Acquiring lock "b39057fd-7acc-437a-bb61-4b84ba4df435-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.668694] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a3dc40-eae6-4d8a-a19d-95b0d9d8ecd1 req-2be7dfde-e739-4b2a-8f63-0b13e239e8cd service nova] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.668869] env[62235]: DEBUG oslo_concurrency.lockutils [req-e3a3dc40-eae6-4d8a-a19d-95b0d9d8ecd1 req-2be7dfde-e739-4b2a-8f63-0b13e239e8cd service nova] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.669057] env[62235]: DEBUG nova.compute.manager [req-e3a3dc40-eae6-4d8a-a19d-95b0d9d8ecd1 req-2be7dfde-e739-4b2a-8f63-0b13e239e8cd service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] No waiting events found dispatching network-vif-plugged-7c34d6b4-c887-439a-8c23-72d06b575f7c {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1179.669237] env[62235]: WARNING nova.compute.manager [req-e3a3dc40-eae6-4d8a-a19d-95b0d9d8ecd1 req-2be7dfde-e739-4b2a-8f63-0b13e239e8cd service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Received unexpected event network-vif-plugged-7c34d6b4-c887-439a-8c23-72d06b575f7c for instance with vm_state building and task_state spawning. [ 1179.756373] env[62235]: DEBUG nova.network.neutron [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Successfully updated port: 7c34d6b4-c887-439a-8c23-72d06b575f7c {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1180.259344] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "refresh_cache-b39057fd-7acc-437a-bb61-4b84ba4df435" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.259551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquired lock "refresh_cache-b39057fd-7acc-437a-bb61-4b84ba4df435" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.259655] env[62235]: DEBUG nova.network.neutron [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1180.797035] env[62235]: DEBUG nova.network.neutron [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1180.926412] env[62235]: DEBUG nova.network.neutron [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Updating instance_info_cache with network_info: [{"id": "7c34d6b4-c887-439a-8c23-72d06b575f7c", "address": "fa:16:3e:66:0b:32", "network": {"id": "32f2cda8-bc05-49b7-bd09-ccf6e7f92ad3", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-550178305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf6f966ac21f49f08bcd0f413f1d6e72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c34d6b4-c8", "ovs_interfaceid": "7c34d6b4-c887-439a-8c23-72d06b575f7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.429493] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Releasing lock "refresh_cache-b39057fd-7acc-437a-bb61-4b84ba4df435" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.429833] env[62235]: DEBUG nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Instance network_info: |[{"id": "7c34d6b4-c887-439a-8c23-72d06b575f7c", "address": "fa:16:3e:66:0b:32", "network": {"id": "32f2cda8-bc05-49b7-bd09-ccf6e7f92ad3", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-550178305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf6f966ac21f49f08bcd0f413f1d6e72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c34d6b4-c8", "ovs_interfaceid": "7c34d6b4-c887-439a-8c23-72d06b575f7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1181.430333] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:0b:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7c34d6b4-c887-439a-8c23-72d06b575f7c', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.437832] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Creating folder: Project (cf6f966ac21f49f08bcd0f413f1d6e72). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1181.438141] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9953d6f5-5d16-4656-bf6d-34f2b9cdb166 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.449041] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Created folder: Project (cf6f966ac21f49f08bcd0f413f1d6e72) in parent group-v273362. [ 1181.449251] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Creating folder: Instances. Parent ref: group-v273535. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1181.449487] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c1ced82-bc23-4eec-af99-974ea540caff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.459133] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Created folder: Instances in parent group-v273535. [ 1181.459350] env[62235]: DEBUG oslo.service.loopingcall [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1181.459538] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1181.459726] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1c11d27-4ab2-43ab-966f-f0bd31e1e5b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.477902] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1181.477902] env[62235]: value = "task-1272299" [ 1181.477902] env[62235]: _type = "Task" [ 1181.477902] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.485024] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272299, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.521643] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1181.521877] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1181.522748] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30a10fe-202b-4d64-ab1c-d83e1517c5e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.538562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203bbd2d-92e3-4d07-acac-f5724a4988d9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.564616] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-6c24557c-9453-4c92-b5bc-4255c395522e/volume-6c24557c-9453-4c92-b5bc-4255c395522e.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.564897] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97d1ad88-71a0-44b2-b2f7-0754889c7fb2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.583998] env[62235]: DEBUG oslo_vmware.api [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1181.583998] env[62235]: value = "task-1272300" [ 1181.583998] env[62235]: _type = "Task" [ 1181.583998] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.591817] env[62235]: DEBUG oslo_vmware.api [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272300, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.704045] env[62235]: DEBUG nova.compute.manager [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Received event network-changed-7c34d6b4-c887-439a-8c23-72d06b575f7c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1181.704232] env[62235]: DEBUG nova.compute.manager [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Refreshing instance network info cache due to event network-changed-7c34d6b4-c887-439a-8c23-72d06b575f7c. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1181.704488] env[62235]: DEBUG oslo_concurrency.lockutils [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] Acquiring lock "refresh_cache-b39057fd-7acc-437a-bb61-4b84ba4df435" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.704659] env[62235]: DEBUG oslo_concurrency.lockutils [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] Acquired lock "refresh_cache-b39057fd-7acc-437a-bb61-4b84ba4df435" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.704854] env[62235]: DEBUG nova.network.neutron [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Refreshing network info cache for port 7c34d6b4-c887-439a-8c23-72d06b575f7c {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1181.987681] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272299, 'name': CreateVM_Task, 'duration_secs': 0.391814} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.987975] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1181.988551] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.988723] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.989056] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1181.989310] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60af88e4-2d3d-4f30-aedf-721ae9286eac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.993556] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1181.993556] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5297e1f8-dc28-89a3-1a70-33d5289d6255" [ 1181.993556] env[62235]: _type = "Task" [ 1181.993556] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.000775] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5297e1f8-dc28-89a3-1a70-33d5289d6255, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.093780] env[62235]: DEBUG oslo_vmware.api [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272300, 'name': ReconfigVM_Task, 'duration_secs': 0.403672} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.094063] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-6c24557c-9453-4c92-b5bc-4255c395522e/volume-6c24557c-9453-4c92-b5bc-4255c395522e.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.098542] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-599bd9e7-424f-4978-bc01-35d69b124ad9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.114153] env[62235]: DEBUG oslo_vmware.api [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1182.114153] env[62235]: value = "task-1272301" [ 1182.114153] env[62235]: _type = "Task" [ 1182.114153] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.121496] env[62235]: DEBUG oslo_vmware.api [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272301, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.387918] env[62235]: DEBUG nova.network.neutron [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Updated VIF entry in instance network info cache for port 7c34d6b4-c887-439a-8c23-72d06b575f7c. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1182.388327] env[62235]: DEBUG nova.network.neutron [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Updating instance_info_cache with network_info: [{"id": "7c34d6b4-c887-439a-8c23-72d06b575f7c", "address": "fa:16:3e:66:0b:32", "network": {"id": "32f2cda8-bc05-49b7-bd09-ccf6e7f92ad3", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-550178305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf6f966ac21f49f08bcd0f413f1d6e72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c34d6b4-c8", "ovs_interfaceid": "7c34d6b4-c887-439a-8c23-72d06b575f7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.503703] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5297e1f8-dc28-89a3-1a70-33d5289d6255, 'name': SearchDatastore_Task, 'duration_secs': 0.010661} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.504008] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.504276] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1182.504518] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.504668] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.504846] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.505119] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77c11eb7-0be6-4653-8a60-274d097c6578 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.512892] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.513082] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1182.513739] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae241b4a-86fe-4a6e-819a-a3b7e4ad7b2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.518409] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1182.518409] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52af7c40-b79b-16d3-2073-8ae63d1463c7" [ 1182.518409] env[62235]: _type = "Task" [ 1182.518409] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.525473] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52af7c40-b79b-16d3-2073-8ae63d1463c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.624977] env[62235]: DEBUG oslo_vmware.api [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272301, 'name': ReconfigVM_Task, 'duration_secs': 0.127524} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.625316] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1182.890809] env[62235]: DEBUG oslo_concurrency.lockutils [req-c9f6c0b2-25ed-4071-b37e-e6458978146f req-598c2715-1875-41d6-8201-7d3ef0e864e4 service nova] Releasing lock "refresh_cache-b39057fd-7acc-437a-bb61-4b84ba4df435" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.028675] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52af7c40-b79b-16d3-2073-8ae63d1463c7, 'name': SearchDatastore_Task, 'duration_secs': 0.007844} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.029497] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75bac984-729f-4393-8865-d9c5244707e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.034511] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1183.034511] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e009f0-22b0-bbde-fe74-2083c942f2a2" [ 1183.034511] env[62235]: _type = "Task" [ 1183.034511] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.041769] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e009f0-22b0-bbde-fe74-2083c942f2a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.544354] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e009f0-22b0-bbde-fe74-2083c942f2a2, 'name': SearchDatastore_Task, 'duration_secs': 0.008703} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.544609] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.544870] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b39057fd-7acc-437a-bb61-4b84ba4df435/b39057fd-7acc-437a-bb61-4b84ba4df435.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1183.545245] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dad61a6c-288d-4944-aadd-a14479592d52 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.552549] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1183.552549] env[62235]: value = "task-1272302" [ 1183.552549] env[62235]: _type = "Task" [ 1183.552549] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.559811] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272302, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.659173] env[62235]: DEBUG nova.objects.instance [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'flavor' on Instance uuid a293951c-183d-4bfe-b41e-2b3408f2eee3 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.063445] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272302, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.403928} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.064223] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] b39057fd-7acc-437a-bb61-4b84ba4df435/b39057fd-7acc-437a-bb61-4b84ba4df435.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1184.064458] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1184.064943] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9eb423ec-d02a-4aee-b2b8-3f2f1f97461c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.071608] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1184.071608] env[62235]: value = "task-1272303" [ 1184.071608] env[62235]: _type = "Task" [ 1184.071608] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.079430] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.164286] env[62235]: DEBUG oslo_concurrency.lockutils [None req-459464a6-7d38-4fea-83c5-3d4190859b19 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.242s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.555588] env[62235]: INFO nova.compute.manager [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Rebuilding instance [ 1184.582711] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058976} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.586297] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1184.587409] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d673649-d367-456a-b846-2d8bfae3d773 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.608883] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] b39057fd-7acc-437a-bb61-4b84ba4df435/b39057fd-7acc-437a-bb61-4b84ba4df435.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1184.611342] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c598fa9-fd1f-4807-b4f6-30c83ba3d2ae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.627536] env[62235]: DEBUG nova.compute.manager [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.628279] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c556da-7ea6-4463-a7c8-e917bcd9b0ab {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.638419] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1184.638419] env[62235]: value = "task-1272304" [ 1184.638419] env[62235]: _type = "Task" [ 1184.638419] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.645489] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272304, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.139293] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.139691] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ddaf3962-6d2f-4e0c-b5e9-4662df9244be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.150950] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272304, 'name': ReconfigVM_Task, 'duration_secs': 0.428354} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.152154] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Reconfigured VM instance instance-0000006b to attach disk [datastore1] b39057fd-7acc-437a-bb61-4b84ba4df435/b39057fd-7acc-437a-bb61-4b84ba4df435.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1185.152780] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1185.152780] env[62235]: value = "task-1272305" [ 1185.152780] env[62235]: _type = "Task" [ 1185.152780] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.152963] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67e8d804-0e68-4a0d-a2c3-be075835f0a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.162133] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1185.162133] env[62235]: value = "task-1272306" [ 1185.162133] env[62235]: _type = "Task" [ 1185.162133] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.164859] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.172432] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272306, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.663959] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272305, 'name': PowerOffVM_Task, 'duration_secs': 0.2395} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.664248] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1185.673788] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272306, 'name': Rename_Task, 'duration_secs': 0.145389} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.674048] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1185.674286] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01cd189b-e420-4136-896e-22337dbccc10 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.680081] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1185.680081] env[62235]: value = "task-1272307" [ 1185.680081] env[62235]: _type = "Task" [ 1185.680081] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.687230] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.718610] env[62235]: INFO nova.compute.manager [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Detaching volume 6c24557c-9453-4c92-b5bc-4255c395522e [ 1185.749157] env[62235]: INFO nova.virt.block_device [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Attempting to driver detach volume 6c24557c-9453-4c92-b5bc-4255c395522e from mountpoint /dev/sdb [ 1185.749543] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1185.749625] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1185.750539] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fb63f3-0600-4fb6-8240-a0feae1292d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.771678] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15ad5ac-4881-49cd-86e3-8f45236382dc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.778194] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941a5cf5-5164-4622-834f-bf7e7a1f4bec {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.798102] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b187fa-da03-46d5-af5a-4b4f6f876233 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.812554] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] The volume has not been displaced from its original location: [datastore1] volume-6c24557c-9453-4c92-b5bc-4255c395522e/volume-6c24557c-9453-4c92-b5bc-4255c395522e.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1185.817884] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1185.818176] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52af76af-3fc6-4549-b640-e8b57788bf3b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.835750] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1185.835750] env[62235]: value = "task-1272308" [ 1185.835750] env[62235]: _type = "Task" [ 1185.835750] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.846790] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272308, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.985702] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.985978] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.986188] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1185.986354] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1186.190069] env[62235]: DEBUG oslo_vmware.api [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272307, 'name': PowerOnVM_Task, 'duration_secs': 0.47335} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.190455] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1186.190553] env[62235]: INFO nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Took 6.74 seconds to spawn the instance on the hypervisor. [ 1186.190738] env[62235]: DEBUG nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.191571] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ae88b4-aec2-434b-a9ee-c520042530bd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.345398] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272308, 'name': ReconfigVM_Task, 'duration_secs': 0.191227} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.345696] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1186.350233] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e8e7331-cdb8-46cf-a61e-910d9a7dbaa5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.365811] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1186.365811] env[62235]: value = "task-1272309" [ 1186.365811] env[62235]: _type = "Task" [ 1186.365811] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.376484] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272309, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.490046] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Skipping network cache update for instance because it is Building. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1186.533470] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.533620] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.533767] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1186.533926] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid 0feb5395-f648-419a-998b-eff941c16464 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.707858] env[62235]: INFO nova.compute.manager [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Took 11.43 seconds to build instance. [ 1186.875601] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272309, 'name': ReconfigVM_Task, 'duration_secs': 0.156237} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.875891] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1186.967734] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "b39057fd-7acc-437a-bb61-4b84ba4df435" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.210548] env[62235]: DEBUG oslo_concurrency.lockutils [None req-55167a45-cf41-4dc5-8159-25bf89a4930d tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.939s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.210955] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.243s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.211046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "b39057fd-7acc-437a-bb61-4b84ba4df435-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.211271] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.211540] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.213860] env[62235]: INFO nova.compute.manager [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Terminating instance [ 1187.215614] env[62235]: DEBUG nova.compute.manager [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1187.215822] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1187.216658] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c73c355-ef2b-4ef7-ab23-4a4d34e332d8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.224220] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.224457] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf551b02-9af7-41ae-96f4-3af2f2914e3c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.230168] env[62235]: DEBUG oslo_vmware.api [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1187.230168] env[62235]: value = "task-1272310" [ 1187.230168] env[62235]: _type = "Task" [ 1187.230168] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.237739] env[62235]: DEBUG oslo_vmware.api [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.741146] env[62235]: DEBUG oslo_vmware.api [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272310, 'name': PowerOffVM_Task, 'duration_secs': 0.198978} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.741383] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1187.741555] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.741822] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb493aeb-603b-4884-885f-867ea583a608 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.803133] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.803393] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.803578] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Deleting the datastore file [datastore1] b39057fd-7acc-437a-bb61-4b84ba4df435 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.803855] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da97f7e6-2ff0-4817-857d-710cd60bd9cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.810369] env[62235]: DEBUG oslo_vmware.api [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for the task: (returnval){ [ 1187.810369] env[62235]: value = "task-1272312" [ 1187.810369] env[62235]: _type = "Task" [ 1187.810369] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.817827] env[62235]: DEBUG oslo_vmware.api [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.923121] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.923424] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-095c3dae-5e93-4c72-b627-4531bce148e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.929864] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1187.929864] env[62235]: value = "task-1272313" [ 1187.929864] env[62235]: _type = "Task" [ 1187.929864] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.937191] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.247577] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.319517] env[62235]: DEBUG oslo_vmware.api [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Task: {'id': task-1272312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146666} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.319774] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.319978] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.320193] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.320382] env[62235]: INFO nova.compute.manager [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1188.320626] env[62235]: DEBUG oslo.service.loopingcall [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1188.320819] env[62235]: DEBUG nova.compute.manager [-] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1188.320912] env[62235]: DEBUG nova.network.neutron [-] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1188.439378] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] VM already powered off {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1188.439689] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1188.439957] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1188.440748] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f970a53-1068-4c68-954a-f72901195066 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.460842] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73b100c-30f2-4471-b717-ae85cfb5f54c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.466554] env[62235]: WARNING nova.virt.vmwareapi.driver [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1188.466824] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1188.467538] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97aaa971-7e9e-43e9-b7a5-1f1e396702be {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.473509] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1188.473725] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa2e72ce-fdaf-4f29-9485-6803466e6d16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.531605] env[62235]: DEBUG nova.compute.manager [req-fd8331b0-48d1-487b-bb2c-906beecb9a42 req-58c7755d-6d45-4324-95cf-2df049b6f740 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Received event network-vif-deleted-7c34d6b4-c887-439a-8c23-72d06b575f7c {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1188.531818] env[62235]: INFO nova.compute.manager [req-fd8331b0-48d1-487b-bb2c-906beecb9a42 req-58c7755d-6d45-4324-95cf-2df049b6f740 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Neutron deleted interface 7c34d6b4-c887-439a-8c23-72d06b575f7c; detaching it from the instance and deleting it from the info cache [ 1188.531982] env[62235]: DEBUG nova.network.neutron [req-fd8331b0-48d1-487b-bb2c-906beecb9a42 req-58c7755d-6d45-4324-95cf-2df049b6f740 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.534964] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1188.535175] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1188.535360] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleting the datastore file [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1188.535789] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e22eefcb-5f55-4b9b-bd72-7ff66f3e6df8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.541547] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1188.541547] env[62235]: value = "task-1272315" [ 1188.541547] env[62235]: _type = "Task" [ 1188.541547] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.549962] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.750377] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.750594] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1188.750798] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.750957] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.751117] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.751270] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.751408] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.751549] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.751677] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1188.751818] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.010356] env[62235]: DEBUG nova.network.neutron [-] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.036841] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb6db286-0386-4c0c-a7bb-193e1077d11e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.050393] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8e5ecc-c1f3-4db0-9f5b-e243a7ad8b57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.065127] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124299} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.065404] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1189.065577] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1189.065753] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1189.072672] env[62235]: DEBUG nova.compute.manager [req-fd8331b0-48d1-487b-bb2c-906beecb9a42 req-58c7755d-6d45-4324-95cf-2df049b6f740 service nova] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Detach interface failed, port_id=7c34d6b4-c887-439a-8c23-72d06b575f7c, reason: Instance b39057fd-7acc-437a-bb61-4b84ba4df435 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1189.254645] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.254976] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.255029] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.255262] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1189.256185] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6defe0f-4ded-439c-b4ee-e4f6a073ce5e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.264731] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb75d3c-2eef-4cdc-8475-8cab1ff16e04 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.278409] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158f9330-9cab-4d92-9a0c-bd77677e109c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.284543] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db389006-5823-45d0-9bef-e19d6bb34302 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.312298] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180672MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1189.312435] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.312621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.513970] env[62235]: INFO nova.compute.manager [-] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Took 1.19 seconds to deallocate network for instance. [ 1189.575145] env[62235]: INFO nova.virt.block_device [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Booting with volume 6c24557c-9453-4c92-b5bc-4255c395522e at /dev/sdb [ 1189.607662] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6eaf6ee-e4b1-4749-98e5-39ef76b6e56e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.616093] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc36dad1-b9ba-4346-b8c4-c6ff3bc27e51 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.638337] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e4a808b-2872-4d08-8eca-acb30e8ed5e3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.645421] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256abbd9-f22b-4bc6-94d7-e0cbd47e523e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.666323] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a522a1a8-05a5-4d7a-9218-e1d49f1eb83f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.671995] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc527304-a05d-49d9-94d8-17d25428f84d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.684335] env[62235]: DEBUG nova.virt.block_device [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updating existing volume attachment record: 85e21997-bd77-4cea-a336-4f3301d5a0bc {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1190.019865] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.337339] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0feb5395-f648-419a-998b-eff941c16464 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.337589] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a293951c-183d-4bfe-b41e-2b3408f2eee3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.337639] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b39057fd-7acc-437a-bb61-4b84ba4df435 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1190.337793] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1190.337932] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1190.382943] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efa787a-0abe-4782-9e20-8a639a3787b0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.390331] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05b5129-deb3-4c5c-b54e-9c84d6e19f90 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.420676] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1662bb5-b0f6-4a4c-a549-3944a8fd7b69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.427312] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e11a752-d831-4a1c-81d9-b0c3720d6a25 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.439993] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.943207] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1191.447627] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1191.447991] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.135s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.448124] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.428s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.448358] env[62235]: DEBUG nova.objects.instance [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lazy-loading 'resources' on Instance uuid b39057fd-7acc-437a-bb61-4b84ba4df435 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.786325] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1191.786583] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1191.786750] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.786939] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1191.787104] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.787265] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1191.787506] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1191.787685] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1191.787858] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1191.788041] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1191.788226] env[62235]: DEBUG nova.virt.hardware [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.789092] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6142e8f-64b4-4cb9-9994-593cc99ee398 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.796850] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98534e7-f6e5-4226-909c-a61edb4750b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.809775] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:34:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b29af1b5-b582-4e06-945e-cadfe1b86830', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.817009] env[62235]: DEBUG oslo.service.loopingcall [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.817251] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.817452] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae2a7e18-092d-402c-9929-d8dbaa42eecd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.836474] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.836474] env[62235]: value = "task-1272316" [ 1191.836474] env[62235]: _type = "Task" [ 1191.836474] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.843319] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272316, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.001718] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40596d27-abc4-4141-af93-8c8c257a2217 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.009848] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd67b2d-c260-43a2-9699-8aa29854c704 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.041298] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dda1c00-47e3-413f-af6b-696462154906 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.048545] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9973b390-313b-4444-ad8c-b3c25f8a70ff {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.061400] env[62235]: DEBUG nova.compute.provider_tree [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.346215] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272316, 'name': CreateVM_Task, 'duration_secs': 0.292035} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.346394] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1192.347083] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.347265] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.347579] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1192.347819] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61706660-54bc-4830-a411-9346d074f309 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.352012] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1192.352012] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f26dcc-e0c2-bcd7-eea7-a0e9a1482d34" [ 1192.352012] env[62235]: _type = "Task" [ 1192.352012] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.358913] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f26dcc-e0c2-bcd7-eea7-a0e9a1482d34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.565008] env[62235]: DEBUG nova.scheduler.client.report [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1192.861942] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52f26dcc-e0c2-bcd7-eea7-a0e9a1482d34, 'name': SearchDatastore_Task, 'duration_secs': 0.009458} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.862296] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.862571] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.862814] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.862965] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.863176] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.863433] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c81a1a2-fdba-4643-acd7-303eb007fe95 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.871927] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.872152] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1192.872802] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb274464-6b03-48f5-88b1-716580fde384 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.877593] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1192.877593] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52973b60-16f7-485d-cd13-73bfe51df464" [ 1192.877593] env[62235]: _type = "Task" [ 1192.877593] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.884956] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52973b60-16f7-485d-cd13-73bfe51df464, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.070332] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.622s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.087134] env[62235]: INFO nova.scheduler.client.report [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Deleted allocations for instance b39057fd-7acc-437a-bb61-4b84ba4df435 [ 1193.388126] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52973b60-16f7-485d-cd13-73bfe51df464, 'name': SearchDatastore_Task, 'duration_secs': 0.007975} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.389219] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72938ba6-4eed-4324-bd88-ed3b70b92fc6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.395254] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1193.395254] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cdabba-14b7-755b-15d9-a33065333083" [ 1193.395254] env[62235]: _type = "Task" [ 1193.395254] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.402468] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cdabba-14b7-755b-15d9-a33065333083, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.594181] env[62235]: DEBUG oslo_concurrency.lockutils [None req-66d55ed5-531d-40f3-a32f-7304b1d38f9c tempest-ServerPasswordTestJSON-1258604948 tempest-ServerPasswordTestJSON-1258604948-project-member] Lock "b39057fd-7acc-437a-bb61-4b84ba4df435" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.383s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.905692] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52cdabba-14b7-755b-15d9-a33065333083, 'name': SearchDatastore_Task, 'duration_secs': 0.00915} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.905950] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.906227] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1193.906482] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a5d4ec4-d0a1-4cd3-bf86-c5a61af743b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.912679] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1193.912679] env[62235]: value = "task-1272317" [ 1193.912679] env[62235]: _type = "Task" [ 1193.912679] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.919703] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.421815] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272317, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427248} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.422125] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1194.422399] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1194.422914] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8dff6fe8-8ed9-43e6-ba56-9f94a6fb30bc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.429015] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1194.429015] env[62235]: value = "task-1272318" [ 1194.429015] env[62235]: _type = "Task" [ 1194.429015] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.436403] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272318, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.938953] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272318, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066939} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.939264] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1194.940222] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c122fa-8065-4a4b-8c02-a5bd91b5c912 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.963109] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1194.963397] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad3f5909-52df-4f64-938b-441d581f5cb4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.982386] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1194.982386] env[62235]: value = "task-1272319" [ 1194.982386] env[62235]: _type = "Task" [ 1194.982386] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.989569] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.492020] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272319, 'name': ReconfigVM_Task, 'duration_secs': 0.302575} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.492426] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfigured VM instance instance-0000006a to attach disk [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3/a293951c-183d-4bfe-b41e-2b3408f2eee3.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1195.493485] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encryption_options': None, 'encryption_format': None, 'device_type': 'disk', 'encrypted': False, 'disk_bus': None, 'size': 0, 'encryption_secret_uuid': None, 'boot_index': 0, 'device_name': '/dev/sda', 'image_id': '02c64327-1eca-4d55-9f01-62fa8b2a2334'}], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sdb', 'attachment_id': '85e21997-bd77-4cea-a336-4f3301d5a0bc', 'device_type': None, 'disk_bus': None, 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'}, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=62235) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1195.493784] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1195.493901] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1195.494706] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8252923-bec6-476b-a910-c0aefffa708c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.509655] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee840dbf-bf01-4f2b-9ff6-dcec8a28c040 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.533134] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-6c24557c-9453-4c92-b5bc-4255c395522e/volume-6c24557c-9453-4c92-b5bc-4255c395522e.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1195.533394] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43ee75c7-48f7-459a-92b4-d9796e46ad09 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.552009] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1195.552009] env[62235]: value = "task-1272320" [ 1195.552009] env[62235]: _type = "Task" [ 1195.552009] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.559184] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272320, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.061898] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272320, 'name': ReconfigVM_Task, 'duration_secs': 0.391336} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.062288] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-6c24557c-9453-4c92-b5bc-4255c395522e/volume-6c24557c-9453-4c92-b5bc-4255c395522e.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1196.066743] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9e0c573-c570-4a34-bc92-a02aa81d139a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.081048] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1196.081048] env[62235]: value = "task-1272321" [ 1196.081048] env[62235]: _type = "Task" [ 1196.081048] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.088347] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272321, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.590987] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272321, 'name': ReconfigVM_Task, 'duration_secs': 0.135718} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.591465] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1196.592036] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87d48f63-e62a-43f3-9c6f-2495e872c048 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.598348] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1196.598348] env[62235]: value = "task-1272322" [ 1196.598348] env[62235]: _type = "Task" [ 1196.598348] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.605283] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272322, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.108321] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272322, 'name': Rename_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.608774] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272322, 'name': Rename_Task, 'duration_secs': 0.705184} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.609080] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1197.609331] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a55aa9e0-b46c-4aa9-85d3-abdf2b2f6f06 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.615578] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1197.615578] env[62235]: value = "task-1272323" [ 1197.615578] env[62235]: _type = "Task" [ 1197.615578] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.623277] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.125286] env[62235]: DEBUG oslo_vmware.api [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272323, 'name': PowerOnVM_Task, 'duration_secs': 0.411901} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.125647] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1198.125765] env[62235]: DEBUG nova.compute.manager [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1198.126516] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cec80c9-16ca-4bc3-8c9e-b91025f4d902 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.641739] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.642158] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.642250] env[62235]: DEBUG nova.objects.instance [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62235) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1199.650215] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5a18bd17-de70-47f9-8470-37def3bbb01c tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.680315] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.680564] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.184302] env[62235]: INFO nova.compute.manager [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Detaching volume 6c24557c-9453-4c92-b5bc-4255c395522e [ 1200.213823] env[62235]: INFO nova.virt.block_device [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Attempting to driver detach volume 6c24557c-9453-4c92-b5bc-4255c395522e from mountpoint /dev/sdb [ 1200.213949] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1200.214114] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1200.214980] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cd121e-c0e2-40fe-9bf0-21f04e2db4c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.235891] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34ab22d-188e-4c12-aedc-805b349e2390 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.242471] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af994b95-d1ed-4000-b17b-9e59c60456a8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.261982] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcac3e47-16f9-4c32-86f6-e0794cd550b5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.275750] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] The volume has not been displaced from its original location: [datastore1] volume-6c24557c-9453-4c92-b5bc-4255c395522e/volume-6c24557c-9453-4c92-b5bc-4255c395522e.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1200.280798] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1200.281057] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e399700-7cda-41ed-92ab-8caa2e957d85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.297689] env[62235]: DEBUG oslo_vmware.api [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1200.297689] env[62235]: value = "task-1272324" [ 1200.297689] env[62235]: _type = "Task" [ 1200.297689] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.304582] env[62235]: DEBUG oslo_vmware.api [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272324, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.806938] env[62235]: DEBUG oslo_vmware.api [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272324, 'name': ReconfigVM_Task, 'duration_secs': 0.239663} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.807310] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1200.811823] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5c24930-b971-41d7-99ce-11c498ae125e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.826253] env[62235]: DEBUG oslo_vmware.api [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1200.826253] env[62235]: value = "task-1272325" [ 1200.826253] env[62235]: _type = "Task" [ 1200.826253] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.833795] env[62235]: DEBUG oslo_vmware.api [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.335872] env[62235]: DEBUG oslo_vmware.api [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272325, 'name': ReconfigVM_Task, 'duration_secs': 0.140641} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.336192] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273534', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'name': 'volume-6c24557c-9453-4c92-b5bc-4255c395522e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a293951c-183d-4bfe-b41e-2b3408f2eee3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c24557c-9453-4c92-b5bc-4255c395522e', 'serial': '6c24557c-9453-4c92-b5bc-4255c395522e'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1201.874295] env[62235]: DEBUG nova.objects.instance [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'flavor' on Instance uuid a293951c-183d-4bfe-b41e-2b3408f2eee3 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1202.882394] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1e08a469-028b-4c80-90fa-df369308133d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.202s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.905736] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.906153] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.906272] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "a293951c-183d-4bfe-b41e-2b3408f2eee3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.906467] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.906643] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.908944] env[62235]: INFO nova.compute.manager [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Terminating instance [ 1203.911029] env[62235]: DEBUG nova.compute.manager [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1203.911259] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1203.912153] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b44aca-6d5a-48c5-a384-a4a30367fe8b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.919392] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1203.919623] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb0ee3e0-7410-4cb4-b3c6-ee815df0ab7a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.925915] env[62235]: DEBUG oslo_vmware.api [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1203.925915] env[62235]: value = "task-1272326" [ 1203.925915] env[62235]: _type = "Task" [ 1203.925915] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.933349] env[62235]: DEBUG oslo_vmware.api [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.436415] env[62235]: DEBUG oslo_vmware.api [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272326, 'name': PowerOffVM_Task, 'duration_secs': 0.161575} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.436690] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1204.436862] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1204.437127] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17fa8f0d-1f29-4412-a3dc-91a33ed06c87 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.494404] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1204.494628] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1204.494803] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleting the datastore file [datastore2] a293951c-183d-4bfe-b41e-2b3408f2eee3 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1204.495085] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98f9ee9b-02ce-465e-a82b-3c6e1a995806 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.501083] env[62235]: DEBUG oslo_vmware.api [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1204.501083] env[62235]: value = "task-1272328" [ 1204.501083] env[62235]: _type = "Task" [ 1204.501083] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.508709] env[62235]: DEBUG oslo_vmware.api [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272328, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.779724] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.779891] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.010943] env[62235]: DEBUG oslo_vmware.api [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272328, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.119538} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.011303] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1205.011410] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1205.011587] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1205.011768] env[62235]: INFO nova.compute.manager [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1205.012023] env[62235]: DEBUG oslo.service.loopingcall [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1205.012248] env[62235]: DEBUG nova.compute.manager [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1205.012350] env[62235]: DEBUG nova.network.neutron [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1205.285926] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.285926] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1205.442248] env[62235]: DEBUG nova.compute.manager [req-284ce12a-65f5-40d0-8e18-50272c0c9338 req-65b75558-f2a9-41e1-b913-60559c2029d2 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Received event network-vif-deleted-b29af1b5-b582-4e06-945e-cadfe1b86830 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1205.442437] env[62235]: INFO nova.compute.manager [req-284ce12a-65f5-40d0-8e18-50272c0c9338 req-65b75558-f2a9-41e1-b913-60559c2029d2 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Neutron deleted interface b29af1b5-b582-4e06-945e-cadfe1b86830; detaching it from the instance and deleting it from the info cache [ 1205.442672] env[62235]: DEBUG nova.network.neutron [req-284ce12a-65f5-40d0-8e18-50272c0c9338 req-65b75558-f2a9-41e1-b913-60559c2029d2 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.789915] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Skipping network cache update for instance because it is being deleted. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9977}} [ 1205.790120] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Didn't find any instances for network info cache update. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1205.790270] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.790429] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.790575] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.790721] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.790861] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.791012] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.791141] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1205.791286] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.920117] env[62235]: DEBUG nova.network.neutron [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.945256] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7592b59c-f6d5-4f92-bd92-7fab5e6dba43 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.954949] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42d3d2e-d743-40b1-a285-e86d7fcf46b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.977876] env[62235]: DEBUG nova.compute.manager [req-284ce12a-65f5-40d0-8e18-50272c0c9338 req-65b75558-f2a9-41e1-b913-60559c2029d2 service nova] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Detach interface failed, port_id=b29af1b5-b582-4e06-945e-cadfe1b86830, reason: Instance a293951c-183d-4bfe-b41e-2b3408f2eee3 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1206.294284] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.294606] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.294693] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.294848] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1206.295752] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2564936c-1a12-4c26-9806-75bef2bd1822 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.303797] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d6be74-0dff-4d8e-b0f9-24618c6434c5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.317374] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08772f6e-29d2-4a49-9a5a-b6fb229a1111 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.323265] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21529041-93e0-47ed-a311-89381326c7dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.350496] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181134MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1206.350623] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.350801] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.422706] env[62235]: INFO nova.compute.manager [-] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Took 1.41 seconds to deallocate network for instance. [ 1206.928746] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.378654] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0feb5395-f648-419a-998b-eff941c16464 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1207.379058] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance a293951c-183d-4bfe-b41e-2b3408f2eee3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1207.379058] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1207.379244] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1207.414779] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79394299-0575-4f09-8db2-506a63cf4776 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.421542] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4f510a-f0d0-4d19-b56a-0378eb309028 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.450242] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce46bc8a-da6c-4441-afd0-7fbb7843e214 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.456587] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03372e53-f5c0-438a-b885-91aaf4031dcc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.468817] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.972345] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.477314] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1208.477764] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.127s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.478066] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.549s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.478399] env[62235]: DEBUG nova.objects.instance [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'resources' on Instance uuid a293951c-183d-4bfe-b41e-2b3408f2eee3 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1209.018618] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60522fc4-dc04-4520-be8a-069e2c00c435 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.025983] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634b5076-ba62-4180-a149-e39513b28285 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.054965] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54659340-b2e4-414d-a6fb-bbbdefe72658 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.061545] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28486814-826a-442f-a2a6-f82a66efda16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.075478] env[62235]: DEBUG nova.compute.provider_tree [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.578412] env[62235]: DEBUG nova.scheduler.client.report [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1210.084073] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.606s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.104588] env[62235]: INFO nova.scheduler.client.report [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted allocations for instance a293951c-183d-4bfe-b41e-2b3408f2eee3 [ 1210.611915] env[62235]: DEBUG oslo_concurrency.lockutils [None req-4bba5598-cdf8-46a3-bb7e-a264fad6b4eb tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "a293951c-183d-4bfe-b41e-2b3408f2eee3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.706s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.856203] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "b7a22574-29f5-423b-b323-ea5eed8a65c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.856479] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.358431] env[62235]: DEBUG nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1213.877339] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.877625] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.879094] env[62235]: INFO nova.compute.claims [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1214.921807] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bedbd9c3-362b-4600-aabf-dfcb41226cc1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.929294] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b34285b-a619-44a0-864d-ccb064751f7e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.957992] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b6736e-617d-47e9-9304-7d10e4c98dba {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.964739] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6293e152-d074-4b08-b876-123976d91c31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.978357] env[62235]: DEBUG nova.compute.provider_tree [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1215.481634] env[62235]: DEBUG nova.scheduler.client.report [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1215.986508] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.109s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.987086] env[62235]: DEBUG nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1216.491659] env[62235]: DEBUG nova.compute.utils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1216.493299] env[62235]: DEBUG nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1216.493431] env[62235]: DEBUG nova.network.neutron [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1216.545878] env[62235]: DEBUG nova.policy [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91b451aa2fe74f6eabee71b0f4f90f0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cd98f2307d24723b3aeb481b535ace9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1216.793692] env[62235]: DEBUG nova.network.neutron [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Successfully created port: 3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1216.997258] env[62235]: DEBUG nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1218.008836] env[62235]: DEBUG nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1218.034216] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1218.034477] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1218.034637] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1218.034825] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1218.034978] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1218.035149] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1218.035373] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1218.035913] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1218.035913] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1218.035913] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1218.036104] env[62235]: DEBUG nova.virt.hardware [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1218.036864] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7102b3-86af-4753-9588-2cf9fe0bd354 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.044620] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94054dc0-ce72-4046-8d6c-6e18f95c29cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.153947] env[62235]: DEBUG nova.compute.manager [req-c6aab1cc-94a8-45e7-9c30-3b8ed1304634 req-4aca0eaf-18bb-4e0c-b207-0997ad379fc8 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Received event network-vif-plugged-3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1218.154201] env[62235]: DEBUG oslo_concurrency.lockutils [req-c6aab1cc-94a8-45e7-9c30-3b8ed1304634 req-4aca0eaf-18bb-4e0c-b207-0997ad379fc8 service nova] Acquiring lock "b7a22574-29f5-423b-b323-ea5eed8a65c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.154416] env[62235]: DEBUG oslo_concurrency.lockutils [req-c6aab1cc-94a8-45e7-9c30-3b8ed1304634 req-4aca0eaf-18bb-4e0c-b207-0997ad379fc8 service nova] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.154683] env[62235]: DEBUG oslo_concurrency.lockutils [req-c6aab1cc-94a8-45e7-9c30-3b8ed1304634 req-4aca0eaf-18bb-4e0c-b207-0997ad379fc8 service nova] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.154790] env[62235]: DEBUG nova.compute.manager [req-c6aab1cc-94a8-45e7-9c30-3b8ed1304634 req-4aca0eaf-18bb-4e0c-b207-0997ad379fc8 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] No waiting events found dispatching network-vif-plugged-3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1218.154962] env[62235]: WARNING nova.compute.manager [req-c6aab1cc-94a8-45e7-9c30-3b8ed1304634 req-4aca0eaf-18bb-4e0c-b207-0997ad379fc8 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Received unexpected event network-vif-plugged-3cb0be04-88f5-48ea-bb89-6c93a10fb61b for instance with vm_state building and task_state spawning. [ 1218.236345] env[62235]: DEBUG nova.network.neutron [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Successfully updated port: 3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1218.738836] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.739046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.739209] env[62235]: DEBUG nova.network.neutron [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1219.277839] env[62235]: DEBUG nova.network.neutron [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1219.399645] env[62235]: DEBUG nova.network.neutron [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Updating instance_info_cache with network_info: [{"id": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "address": "fa:16:3e:b0:4a:06", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb0be04-88", "ovs_interfaceid": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.902652] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.902983] env[62235]: DEBUG nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Instance network_info: |[{"id": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "address": "fa:16:3e:b0:4a:06", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb0be04-88", "ovs_interfaceid": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1219.903467] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:4a:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cb0be04-88f5-48ea-bb89-6c93a10fb61b', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1219.910792] env[62235]: DEBUG oslo.service.loopingcall [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1219.911000] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1219.911237] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-517bbeea-4f49-4691-b1c1-d18f783e7633 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.932287] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1219.932287] env[62235]: value = "task-1272329" [ 1219.932287] env[62235]: _type = "Task" [ 1219.932287] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.939452] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272329, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.184927] env[62235]: DEBUG nova.compute.manager [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Received event network-changed-3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1220.185215] env[62235]: DEBUG nova.compute.manager [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Refreshing instance network info cache due to event network-changed-3cb0be04-88f5-48ea-bb89-6c93a10fb61b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1220.185502] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] Acquiring lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.185744] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] Acquired lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.185996] env[62235]: DEBUG nova.network.neutron [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Refreshing network info cache for port 3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1220.442356] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272329, 'name': CreateVM_Task, 'duration_secs': 0.298545} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.442682] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1220.443172] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.443372] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.443699] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1220.443953] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdfbd3a2-015a-4130-80b4-8e889412d22d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.448206] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1220.448206] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526b99c8-fc12-a2bd-1d5d-c09695329cd7" [ 1220.448206] env[62235]: _type = "Task" [ 1220.448206] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.455702] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526b99c8-fc12-a2bd-1d5d-c09695329cd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.865502] env[62235]: DEBUG nova.network.neutron [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Updated VIF entry in instance network info cache for port 3cb0be04-88f5-48ea-bb89-6c93a10fb61b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1220.865857] env[62235]: DEBUG nova.network.neutron [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Updating instance_info_cache with network_info: [{"id": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "address": "fa:16:3e:b0:4a:06", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb0be04-88", "ovs_interfaceid": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.957865] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]526b99c8-fc12-a2bd-1d5d-c09695329cd7, 'name': SearchDatastore_Task, 'duration_secs': 0.010359} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.958182] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.958433] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1220.958673] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.958823] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.959026] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1220.959289] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-52d2958d-16b7-4ed2-ba0e-cf0cf2e1284e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.967700] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1220.967893] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1220.968626] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e521227e-7738-41c5-bf87-fa93b3ac49a0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.974127] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1220.974127] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c07948-aaab-ae95-9060-706faf267e02" [ 1220.974127] env[62235]: _type = "Task" [ 1220.974127] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.981797] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c07948-aaab-ae95-9060-706faf267e02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.368443] env[62235]: DEBUG oslo_concurrency.lockutils [req-1b487aab-c52c-478a-842f-a9e035e3ec76 req-e89c2787-d64b-4025-a5f5-baa852ccfb74 service nova] Releasing lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.484943] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c07948-aaab-ae95-9060-706faf267e02, 'name': SearchDatastore_Task, 'duration_secs': 0.0082} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.485724] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf1565ca-1bfd-4e39-88d8-8b5554f089c3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.490423] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1221.490423] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5274b2a9-43dd-f8d1-2f60-d76226a06ca0" [ 1221.490423] env[62235]: _type = "Task" [ 1221.490423] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.497605] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5274b2a9-43dd-f8d1-2f60-d76226a06ca0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.000261] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5274b2a9-43dd-f8d1-2f60-d76226a06ca0, 'name': SearchDatastore_Task, 'duration_secs': 0.008347} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.000538] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.000800] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] b7a22574-29f5-423b-b323-ea5eed8a65c4/b7a22574-29f5-423b-b323-ea5eed8a65c4.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1222.001067] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a5a9d91-20ae-4ea4-8f0c-855a70b2b42a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.008155] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1222.008155] env[62235]: value = "task-1272330" [ 1222.008155] env[62235]: _type = "Task" [ 1222.008155] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.015473] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272330, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.519528] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272330, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.409366} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.519797] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] b7a22574-29f5-423b-b323-ea5eed8a65c4/b7a22574-29f5-423b-b323-ea5eed8a65c4.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1222.520029] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1222.520279] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2d07f6e-a9b2-46d5-9b0d-9018407f21e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.527009] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1222.527009] env[62235]: value = "task-1272331" [ 1222.527009] env[62235]: _type = "Task" [ 1222.527009] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.534757] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272331, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.036542] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272331, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061858} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.036838] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1223.037591] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8d88b7-fcce-4ba0-a777-cb67ee8e2062 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.058033] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] b7a22574-29f5-423b-b323-ea5eed8a65c4/b7a22574-29f5-423b-b323-ea5eed8a65c4.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1223.058264] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-660fcadc-fbd8-4118-b0a6-a1c659f9372b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.078096] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1223.078096] env[62235]: value = "task-1272332" [ 1223.078096] env[62235]: _type = "Task" [ 1223.078096] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.085330] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272332, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.587984] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272332, 'name': ReconfigVM_Task, 'duration_secs': 0.267982} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.588415] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Reconfigured VM instance instance-0000006c to attach disk [datastore2] b7a22574-29f5-423b-b323-ea5eed8a65c4/b7a22574-29f5-423b-b323-ea5eed8a65c4.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1223.588948] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee65590e-ce55-4576-a4b7-6783c7318f7d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.595674] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1223.595674] env[62235]: value = "task-1272333" [ 1223.595674] env[62235]: _type = "Task" [ 1223.595674] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.602902] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272333, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.105619] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272333, 'name': Rename_Task, 'duration_secs': 0.12874} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.105896] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1224.106166] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed0e84fb-8f6b-4450-b72d-e0fa8da913a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.112694] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1224.112694] env[62235]: value = "task-1272334" [ 1224.112694] env[62235]: _type = "Task" [ 1224.112694] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.119829] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.622294] env[62235]: DEBUG oslo_vmware.api [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272334, 'name': PowerOnVM_Task, 'duration_secs': 0.439307} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.622654] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1224.622776] env[62235]: INFO nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Took 6.61 seconds to spawn the instance on the hypervisor. [ 1224.622953] env[62235]: DEBUG nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1224.623733] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faff9965-3b15-4416-97b5-8e4e942dd998 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.140929] env[62235]: INFO nova.compute.manager [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Took 11.28 seconds to build instance. [ 1225.642736] env[62235]: DEBUG oslo_concurrency.lockutils [None req-18ed1b6d-478f-461d-8461-37043e104272 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.786s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.736170] env[62235]: DEBUG nova.compute.manager [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Received event network-changed-3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1225.736723] env[62235]: DEBUG nova.compute.manager [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Refreshing instance network info cache due to event network-changed-3cb0be04-88f5-48ea-bb89-6c93a10fb61b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1225.736723] env[62235]: DEBUG oslo_concurrency.lockutils [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] Acquiring lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.736863] env[62235]: DEBUG oslo_concurrency.lockutils [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] Acquired lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.736971] env[62235]: DEBUG nova.network.neutron [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Refreshing network info cache for port 3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1226.439208] env[62235]: DEBUG nova.network.neutron [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Updated VIF entry in instance network info cache for port 3cb0be04-88f5-48ea-bb89-6c93a10fb61b. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1226.439650] env[62235]: DEBUG nova.network.neutron [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Updating instance_info_cache with network_info: [{"id": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "address": "fa:16:3e:b0:4a:06", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb0be04-88", "ovs_interfaceid": "3cb0be04-88f5-48ea-bb89-6c93a10fb61b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.943184] env[62235]: DEBUG oslo_concurrency.lockutils [req-f63d693d-9a02-470f-8f5d-91f3633ca539 req-fc9818d0-8640-4234-8b96-d336e7bd646d service nova] Releasing lock "refresh_cache-b7a22574-29f5-423b-b323-ea5eed8a65c4" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.048900] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "ba3567eb-9c7c-489a-bb85-eddef758b237" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1245.049882] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.552604] env[62235]: DEBUG nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1246.075581] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.075882] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.077771] env[62235]: INFO nova.compute.claims [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1247.138913] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893a950c-6712-4604-ac0e-ae15952dd5de {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.146840] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd9a98f-130e-420d-a9b1-00a8359e756a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.177960] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52292469-d01f-40db-9ec1-b79ab0122919 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.185305] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1a33a4-b475-461a-b960-448abe039e16 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.199300] env[62235]: DEBUG nova.compute.provider_tree [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1247.702419] env[62235]: DEBUG nova.scheduler.client.report [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1248.207179] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.131s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.207614] env[62235]: DEBUG nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1248.712670] env[62235]: DEBUG nova.compute.utils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1248.714114] env[62235]: DEBUG nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1248.714287] env[62235]: DEBUG nova.network.neutron [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1248.764980] env[62235]: DEBUG nova.policy [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91b451aa2fe74f6eabee71b0f4f90f0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cd98f2307d24723b3aeb481b535ace9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1249.039698] env[62235]: DEBUG nova.network.neutron [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Successfully created port: 773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1249.217420] env[62235]: DEBUG nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1249.722497] env[62235]: INFO nova.virt.block_device [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Booting with volume 97daf10b-3b9b-4cfc-a63d-44d617c3e098 at /dev/sda [ 1249.759715] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0f3cffe-07b4-415b-a2c7-47625535accb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.768514] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f576f6-8a00-41bf-a57f-6449b3122ce2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.792085] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdddc081-b895-48e6-a28a-818584094923 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.799852] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7628771-bb60-447a-b3db-392f08f13c6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.823577] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bc0a1c-9c88-484e-9e2f-7fc9ad212681 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.829796] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b14081-d967-4422-a527-732a30fa2379 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.841841] env[62235]: DEBUG nova.virt.block_device [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating existing volume attachment record: 7a76d663-5f4d-412b-b1e0-5c70129728dd {{(pid=62235) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1250.398089] env[62235]: DEBUG nova.compute.manager [req-1fa7b4ec-62c0-41a7-8304-58ce26a685c9 req-345b0208-1ab1-42f3-bf75-356be934313e service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Received event network-vif-plugged-773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1250.398357] env[62235]: DEBUG oslo_concurrency.lockutils [req-1fa7b4ec-62c0-41a7-8304-58ce26a685c9 req-345b0208-1ab1-42f3-bf75-356be934313e service nova] Acquiring lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.398479] env[62235]: DEBUG oslo_concurrency.lockutils [req-1fa7b4ec-62c0-41a7-8304-58ce26a685c9 req-345b0208-1ab1-42f3-bf75-356be934313e service nova] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.398650] env[62235]: DEBUG oslo_concurrency.lockutils [req-1fa7b4ec-62c0-41a7-8304-58ce26a685c9 req-345b0208-1ab1-42f3-bf75-356be934313e service nova] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1250.398820] env[62235]: DEBUG nova.compute.manager [req-1fa7b4ec-62c0-41a7-8304-58ce26a685c9 req-345b0208-1ab1-42f3-bf75-356be934313e service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] No waiting events found dispatching network-vif-plugged-773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1250.399078] env[62235]: WARNING nova.compute.manager [req-1fa7b4ec-62c0-41a7-8304-58ce26a685c9 req-345b0208-1ab1-42f3-bf75-356be934313e service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Received unexpected event network-vif-plugged-773c38c4-89d2-4ca2-8d65-59200adf41ff for instance with vm_state building and task_state block_device_mapping. [ 1250.481688] env[62235]: DEBUG nova.network.neutron [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Successfully updated port: 773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1250.983963] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.984136] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.984280] env[62235]: DEBUG nova.network.neutron [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1251.514596] env[62235]: DEBUG nova.network.neutron [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1251.639935] env[62235]: DEBUG nova.network.neutron [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [{"id": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "address": "fa:16:3e:0e:f0:b0", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773c38c4-89", "ovs_interfaceid": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.918453] env[62235]: DEBUG nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1251.918988] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1251.919222] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1251.919384] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1251.919573] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1251.919726] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1251.919878] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1251.920098] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1251.920268] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1251.920488] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1251.920664] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1251.920842] env[62235]: DEBUG nova.virt.hardware [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1251.921709] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0668b8bc-c9e1-4d0e-b8eb-1e8bea71b253 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.929529] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f02aa28-f6aa-4ee3-ad76-d2c873cf3687 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.143133] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1252.143573] env[62235]: DEBUG nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Instance network_info: |[{"id": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "address": "fa:16:3e:0e:f0:b0", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773c38c4-89", "ovs_interfaceid": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1252.144036] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:f0:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '773c38c4-89d2-4ca2-8d65-59200adf41ff', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1252.151332] env[62235]: DEBUG oslo.service.loopingcall [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1252.151553] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1252.151772] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-696cfe7c-ae61-441d-947b-5ae89e5b858a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.171648] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1252.171648] env[62235]: value = "task-1272343" [ 1252.171648] env[62235]: _type = "Task" [ 1252.171648] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.178956] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272343, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.423735] env[62235]: DEBUG nova.compute.manager [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Received event network-changed-773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1252.423949] env[62235]: DEBUG nova.compute.manager [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Refreshing instance network info cache due to event network-changed-773c38c4-89d2-4ca2-8d65-59200adf41ff. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1252.424188] env[62235]: DEBUG oslo_concurrency.lockutils [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] Acquiring lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1252.424356] env[62235]: DEBUG oslo_concurrency.lockutils [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] Acquired lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.424512] env[62235]: DEBUG nova.network.neutron [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Refreshing network info cache for port 773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1252.681921] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272343, 'name': CreateVM_Task, 'duration_secs': 0.311325} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.682233] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1252.682764] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'mount_device': '/dev/sda', 'attachment_id': '7a76d663-5f4d-412b-b1e0-5c70129728dd', 'device_type': None, 'disk_bus': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273541', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'name': 'volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba3567eb-9c7c-489a-bb85-eddef758b237', 'attached_at': '', 'detached_at': '', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'serial': '97daf10b-3b9b-4cfc-a63d-44d617c3e098'}, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=62235) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1252.682982] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Root volume attach. Driver type: vmdk {{(pid=62235) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1252.683726] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac7bb71-f79e-4cea-88b7-71dedbf1ab40 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.690627] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb5ea3b-2cc4-462b-afff-e8bffc4ee207 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.696443] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645f8257-dc95-4d2a-a820-9da74b9556e6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.701679] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-2b333bc6-44d2-4173-83a9-c2f8a1b4c573 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.708841] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1252.708841] env[62235]: value = "task-1272344" [ 1252.708841] env[62235]: _type = "Task" [ 1252.708841] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.715958] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272344, 'name': RelocateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.209504] env[62235]: DEBUG nova.network.neutron [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updated VIF entry in instance network info cache for port 773c38c4-89d2-4ca2-8d65-59200adf41ff. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1253.209903] env[62235]: DEBUG nova.network.neutron [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [{"id": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "address": "fa:16:3e:0e:f0:b0", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773c38c4-89", "ovs_interfaceid": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.222832] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272344, 'name': RelocateVM_Task} progress is 43%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.717783] env[62235]: DEBUG oslo_concurrency.lockutils [req-18e90c26-e686-4cd1-9e00-54d412b9a0e6 req-e6543ce7-3e4a-42b8-83d5-b37e7d6ddf75 service nova] Releasing lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1253.721704] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272344, 'name': RelocateVM_Task} progress is 58%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.222810] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272344, 'name': RelocateVM_Task} progress is 75%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.723088] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272344, 'name': RelocateVM_Task} progress is 89%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.224778] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272344, 'name': RelocateVM_Task} progress is 97%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.724910] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272344, 'name': RelocateVM_Task, 'duration_secs': 2.837707} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.725211] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Volume attach. Driver type: vmdk {{(pid=62235) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1255.725416] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273541', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'name': 'volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba3567eb-9c7c-489a-bb85-eddef758b237', 'attached_at': '', 'detached_at': '', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'serial': '97daf10b-3b9b-4cfc-a63d-44d617c3e098'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1255.726163] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361ce1cc-544c-4dce-811e-a9982ec85e0a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.741015] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4557127f-c297-4c70-95b3-c5f20e2cdbf9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.762971] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098/volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1255.763231] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8af3eb9-198a-4326-aaab-31870bd100ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.781494] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1255.781494] env[62235]: value = "task-1272345" [ 1255.781494] env[62235]: _type = "Task" [ 1255.781494] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.789840] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272345, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.291384] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.791356] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272345, 'name': ReconfigVM_Task, 'duration_secs': 0.827968} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.791683] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfigured VM instance instance-0000006d to attach disk [datastore2] volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098/volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1256.796160] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f84dec58-fa79-4e28-9736-e55854be8d57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.810211] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1256.810211] env[62235]: value = "task-1272346" [ 1256.810211] env[62235]: _type = "Task" [ 1256.810211] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.817614] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272346, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.320471] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272346, 'name': ReconfigVM_Task, 'duration_secs': 0.113881} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.320667] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273541', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'name': 'volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba3567eb-9c7c-489a-bb85-eddef758b237', 'attached_at': '', 'detached_at': '', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'serial': '97daf10b-3b9b-4cfc-a63d-44d617c3e098'} {{(pid=62235) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1257.321165] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b33c4b06-8a4e-47ea-9021-bbe368683523 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.327835] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1257.327835] env[62235]: value = "task-1272347" [ 1257.327835] env[62235]: _type = "Task" [ 1257.327835] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.335861] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272347, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.837499] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272347, 'name': Rename_Task, 'duration_secs': 0.18193} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.837856] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1257.838238] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4c063b5-a891-41d7-9063-1f241e5b1b18 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.844556] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1257.844556] env[62235]: value = "task-1272348" [ 1257.844556] env[62235]: _type = "Task" [ 1257.844556] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.854069] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272348, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.354769] env[62235]: DEBUG oslo_vmware.api [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272348, 'name': PowerOnVM_Task, 'duration_secs': 0.422448} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.355067] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1258.355277] env[62235]: INFO nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Took 6.44 seconds to spawn the instance on the hypervisor. [ 1258.355458] env[62235]: DEBUG nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1258.356221] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8343f96c-0f60-488d-a9eb-1093a58993fe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.874697] env[62235]: INFO nova.compute.manager [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Took 12.82 seconds to build instance. [ 1259.376942] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5f0dfd45-84fc-47ff-96d2-fbc0109fa97e tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.328s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.575025] env[62235]: DEBUG nova.compute.manager [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Received event network-changed-8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1259.575244] env[62235]: DEBUG nova.compute.manager [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Refreshing instance network info cache due to event network-changed-8e87c7f6-11e9-49d5-8326-57e2140c176f. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1259.575473] env[62235]: DEBUG oslo_concurrency.lockutils [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] Acquiring lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1259.575619] env[62235]: DEBUG oslo_concurrency.lockutils [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] Acquired lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.575786] env[62235]: DEBUG nova.network.neutron [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Refreshing network info cache for port 8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1260.277615] env[62235]: DEBUG nova.network.neutron [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updated VIF entry in instance network info cache for port 8e87c7f6-11e9-49d5-8326-57e2140c176f. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1260.277974] env[62235]: DEBUG nova.network.neutron [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1260.780768] env[62235]: DEBUG oslo_concurrency.lockutils [req-d8565ff7-af18-457d-b94c-18c3b877fe86 req-e8f83f92-8e6f-4b61-9154-d5c3b7e3d132 service nova] Releasing lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1261.000691] env[62235]: DEBUG nova.compute.manager [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Stashing vm_state: active {{(pid=62235) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1261.521022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.521022] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.599235] env[62235]: DEBUG nova.compute.manager [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Received event network-changed-773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1261.599448] env[62235]: DEBUG nova.compute.manager [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Refreshing instance network info cache due to event network-changed-773c38c4-89d2-4ca2-8d65-59200adf41ff. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1261.599667] env[62235]: DEBUG oslo_concurrency.lockutils [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] Acquiring lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1261.599814] env[62235]: DEBUG oslo_concurrency.lockutils [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] Acquired lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1261.599980] env[62235]: DEBUG nova.network.neutron [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Refreshing network info cache for port 773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1262.025730] env[62235]: INFO nova.compute.claims [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1262.300496] env[62235]: DEBUG nova.network.neutron [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updated VIF entry in instance network info cache for port 773c38c4-89d2-4ca2-8d65-59200adf41ff. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1262.300496] env[62235]: DEBUG nova.network.neutron [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [{"id": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "address": "fa:16:3e:0e:f0:b0", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773c38c4-89", "ovs_interfaceid": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.531722] env[62235]: INFO nova.compute.resource_tracker [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating resource usage from migration 750458f7-2ab3-4727-900a-2164338d9f13 [ 1262.587266] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6041a54b-0f3e-4059-8804-a90899f4cc8f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.595086] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c13d80e-730d-41dd-9d4a-dcd8231348f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.624809] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede96534-35b6-4872-bc68-7ef3400653e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.631807] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008c9eb4-8b69-49ec-85fe-834571ecf284 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.644393] env[62235]: DEBUG nova.compute.provider_tree [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1262.803017] env[62235]: DEBUG oslo_concurrency.lockutils [req-df677ff1-7b6a-484a-aa82-fdc3a558e709 req-65cdf46b-7d62-4fe1-acee-3a2926c52240 service nova] Releasing lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1263.147141] env[62235]: DEBUG nova.scheduler.client.report [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1263.651830] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.131s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.652185] env[62235]: INFO nova.compute.manager [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Migrating [ 1264.166354] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.166550] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.166739] env[62235]: DEBUG nova.network.neutron [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1264.860892] env[62235]: DEBUG nova.network.neutron [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [{"id": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "address": "fa:16:3e:0e:f0:b0", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773c38c4-89", "ovs_interfaceid": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.363890] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1266.879163] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6c0941-4289-4c5a-aa9b-c6d37039fa31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.898536] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance 'ba3567eb-9c7c-489a-bb85-eddef758b237' progress to 0 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1267.404669] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1267.404987] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6635a78c-fb14-4fd0-bce3-b9909db3afd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.411941] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1267.411941] env[62235]: value = "task-1272349" [ 1267.411941] env[62235]: _type = "Task" [ 1267.411941] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.420042] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272349, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.921797] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272349, 'name': PowerOffVM_Task, 'duration_secs': 0.174557} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.922088] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1267.922280] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance 'ba3567eb-9c7c-489a-bb85-eddef758b237' progress to 17 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1268.428863] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1268.429129] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1268.429298] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1268.429484] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1268.429668] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1268.429794] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1268.430013] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1268.430187] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1268.430361] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1268.430547] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1268.430738] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1268.435752] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4aa995b4-e4ab-47a6-95ec-edbb52f8ace4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.452013] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1268.452013] env[62235]: value = "task-1272350" [ 1268.452013] env[62235]: _type = "Task" [ 1268.452013] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.459792] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272350, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.479872] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.480156] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.480348] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1268.480566] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1268.962051] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272350, 'name': ReconfigVM_Task, 'duration_secs': 0.145109} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.962587] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance 'ba3567eb-9c7c-489a-bb85-eddef758b237' progress to 33 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1269.010381] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1269.010552] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.010700] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1269.010854] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid 0feb5395-f648-419a-998b-eff941c16464 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1269.468831] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1269.469230] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1269.469481] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1269.469758] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1269.469996] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1269.470248] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1269.470573] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1269.470820] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1269.471096] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1269.471358] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1269.471627] env[62235]: DEBUG nova.virt.hardware [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1269.477012] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfiguring VM instance instance-0000006d to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1269.477356] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41c02482-397d-416a-80b0-365400403256 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.497208] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1269.497208] env[62235]: value = "task-1272351" [ 1269.497208] env[62235]: _type = "Task" [ 1269.497208] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.504930] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.007128] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272351, 'name': ReconfigVM_Task, 'duration_secs': 0.14674} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.007418] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfigured VM instance instance-0000006d to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1270.008170] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b54756f-4099-479a-8b86-0b13482fee85 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.030740] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098/volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1270.030977] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b02880fa-02d6-477d-abb2-a31cdcce9c73 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.050401] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1270.050401] env[62235]: value = "task-1272352" [ 1270.050401] env[62235]: _type = "Task" [ 1270.050401] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.059181] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272352, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.560499] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272352, 'name': ReconfigVM_Task, 'duration_secs': 0.232014} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.560777] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfigured VM instance instance-0000006d to attach disk [datastore2] volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098/volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098.vmdk or device None with type thin {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1270.561034] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance 'ba3567eb-9c7c-489a-bb85-eddef758b237' progress to 50 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1270.725211] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [{"id": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "address": "fa:16:3e:d3:70:5a", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e87c7f6-11", "ovs_interfaceid": "8e87c7f6-11e9-49d5-8326-57e2140c176f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.068226] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd60c4a-9ebc-4e28-85f7-626baf777506 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.087101] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ef3d9e-7ff3-40c9-9465-f63b2bf3ccbb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.104482] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance 'ba3567eb-9c7c-489a-bb85-eddef758b237' progress to 67 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1271.228276] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-0feb5395-f648-419a-998b-eff941c16464" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1271.228435] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1271.228613] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.228771] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.228919] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.229083] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.229231] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.229378] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.229507] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1271.229652] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.732759] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.733021] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.733196] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.733352] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1271.734253] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d452260-485b-49d7-b597-1a5ba4c1606d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.743209] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485575e0-6594-4ec5-a7e7-14616672c12e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.756526] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a629ffed-1b2b-4aa5-8d19-6176f64af76f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.762445] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654b3115-a632-41b9-bfda-4af58a325ad7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.791204] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181327MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1271.791341] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.791522] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.780575] env[62235]: DEBUG nova.network.neutron [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Port 773c38c4-89d2-4ca2-8d65-59200adf41ff binding to destination host cpu-1 is already ACTIVE {{(pid=62235) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1272.798721] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Applying migration context for instance ba3567eb-9c7c-489a-bb85-eddef758b237 as it has an incoming, in-progress migration 750458f7-2ab3-4727-900a-2164338d9f13. Migration status is migrating {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1272.799422] env[62235]: INFO nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating resource usage from migration 750458f7-2ab3-4727-900a-2164338d9f13 [ 1272.818428] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0feb5395-f648-419a-998b-eff941c16464 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1272.818581] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance b7a22574-29f5-423b-b323-ea5eed8a65c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1272.818709] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Migration 750458f7-2ab3-4727-900a-2164338d9f13 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1272.818829] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ba3567eb-9c7c-489a-bb85-eddef758b237 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1272.819008] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1272.819155] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1272.872528] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b38e4a-f799-436d-bb71-be5fd989726c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.880089] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f086aaad-3e67-40f7-8cfb-e118a2fad4e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.910498] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4786d80-ecf9-42b8-a4a6-8e854e8a5915 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.917244] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd85090-2094-406d-85f4-9c780688f045 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.929667] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1273.432597] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1273.762526] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.762711] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.762894] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.936615] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1273.936979] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.145s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.796973] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.797183] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.797368] env[62235]: DEBUG nova.network.neutron [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1275.507172] env[62235]: DEBUG nova.network.neutron [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [{"id": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "address": "fa:16:3e:0e:f0:b0", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773c38c4-89", "ovs_interfaceid": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1276.010072] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1276.522089] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d28eac-d4e8-40f0-8798-fedf7619dc72 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.529176] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6319380-d70b-494f-bcd5-e8b5baf5d28f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.626160] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad7af69-108f-4c54-bdac-50953a0fb8a3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.645267] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93a6e4e-8260-4c5d-8581-9a039eddca31 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.651848] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance 'ba3567eb-9c7c-489a-bb85-eddef758b237' progress to 83 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1278.158622] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1278.158933] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3def5ce4-e374-4554-a879-67c1b4769e8c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.167954] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1278.167954] env[62235]: value = "task-1272353" [ 1278.167954] env[62235]: _type = "Task" [ 1278.167954] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.175197] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272353, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.678066] env[62235]: DEBUG oslo_vmware.api [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272353, 'name': PowerOnVM_Task, 'duration_secs': 0.356789} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.678460] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1278.678555] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ae40ea34-9f55-4f40-b6fe-e035834d2dc6 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance 'ba3567eb-9c7c-489a-bb85-eddef758b237' progress to 100 {{(pid=62235) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1280.444050] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "ba3567eb-9c7c-489a-bb85-eddef758b237" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.444050] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.444473] env[62235]: DEBUG nova.compute.manager [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Going to confirm migration 6 {{(pid=62235) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1280.980441] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.980666] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquired lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.980860] env[62235]: DEBUG nova.network.neutron [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1280.981067] env[62235]: DEBUG nova.objects.instance [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'info_cache' on Instance uuid ba3567eb-9c7c-489a-bb85-eddef758b237 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1282.180370] env[62235]: DEBUG nova.network.neutron [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [{"id": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "address": "fa:16:3e:0e:f0:b0", "network": {"id": "9bd44e0c-75c4-4cad-a46e-5e6372ef101f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1926894705-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cd98f2307d24723b3aeb481b535ace9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773c38c4-89", "ovs_interfaceid": "773c38c4-89d2-4ca2-8d65-59200adf41ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1282.683656] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Releasing lock "refresh_cache-ba3567eb-9c7c-489a-bb85-eddef758b237" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1282.683968] env[62235]: DEBUG nova.objects.instance [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'migration_context' on Instance uuid ba3567eb-9c7c-489a-bb85-eddef758b237 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1283.186951] env[62235]: DEBUG nova.objects.base [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62235) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1283.187869] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162e3e03-d1ad-49c4-9dda-a94c3bd308c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.206971] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b6c357e-0cb6-405e-8cd6-4e88be1c597e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.212306] env[62235]: DEBUG oslo_vmware.api [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1283.212306] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d80eb1-88a1-7f3f-2404-b05655e0c812" [ 1283.212306] env[62235]: _type = "Task" [ 1283.212306] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.219735] env[62235]: DEBUG oslo_vmware.api [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d80eb1-88a1-7f3f-2404-b05655e0c812, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.722355] env[62235]: DEBUG oslo_vmware.api [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52d80eb1-88a1-7f3f-2404-b05655e0c812, 'name': SearchDatastore_Task, 'duration_secs': 0.008217} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.722695] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1283.722921] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1284.282362] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2c2509-c0e0-4d19-93db-c3e0ea69cf07 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.289597] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9a7688-380b-4b8f-80f8-e7e251b89e19 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.319834] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81626fb3-949f-4093-9b3e-a438923f41fb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.326605] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2fd5e9-6c05-4e0f-801f-375e76b8f1b2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.339306] env[62235]: DEBUG nova.compute.provider_tree [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1284.842818] env[62235]: DEBUG nova.scheduler.client.report [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1285.853172] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.130s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.410860] env[62235]: INFO nova.scheduler.client.report [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted allocation for migration 750458f7-2ab3-4727-900a-2164338d9f13 [ 1286.819980] env[62235]: INFO nova.compute.manager [None req-24366818-9fac-4c03-8654-22c54ee0c794 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Get console output [ 1286.820376] env[62235]: WARNING nova.virt.vmwareapi.driver [None req-24366818-9fac-4c03-8654-22c54ee0c794 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] The console log is missing. Check your VSPC configuration [ 1286.916260] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d5cfc9b-21cf-4c41-8254-0b27cc09ae91 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.472s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.768657] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "ba3567eb-9c7c-489a-bb85-eddef758b237" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.769077] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.769255] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.769461] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.769639] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.772827] env[62235]: INFO nova.compute.manager [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Terminating instance [ 1313.774758] env[62235]: DEBUG nova.compute.manager [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1313.774972] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1313.775273] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e12a67e6-879c-42d0-8d9a-8120fd228fbb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.783112] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1313.783112] env[62235]: value = "task-1272354" [ 1313.783112] env[62235]: _type = "Task" [ 1313.783112] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.790868] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272354, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.292679] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272354, 'name': PowerOffVM_Task, 'duration_secs': 0.180856} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.292935] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1314.293185] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Volume detach. Driver type: vmdk {{(pid=62235) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1314.294197] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273541', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'name': 'volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ba3567eb-9c7c-489a-bb85-eddef758b237', 'attached_at': '2024-10-12T06:31:01.000000', 'detached_at': '', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'serial': '97daf10b-3b9b-4cfc-a63d-44d617c3e098'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1314.294197] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ab1cab-832e-4d56-985f-cf06a7bd2e62 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.311904] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9fadfe-27a2-4bec-85ec-b66e67829ab2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.317923] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84519707-0856-4245-9e12-5466ffd86e24 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.334575] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1195fd-76db-4243-9e80-d4d48ab9701f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.348119] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] The volume has not been displaced from its original location: [datastore2] volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098/volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098.vmdk. No consolidation needed. {{(pid=62235) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1314.353277] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfiguring VM instance instance-0000006d to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1314.353551] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a674b90-3573-4fef-ad84-509478709fc4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.370235] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1314.370235] env[62235]: value = "task-1272355" [ 1314.370235] env[62235]: _type = "Task" [ 1314.370235] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.378568] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272355, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.879995] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272355, 'name': ReconfigVM_Task, 'duration_secs': 0.139856} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.880513] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Reconfigured VM instance instance-0000006d to detach disk 2000 {{(pid=62235) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1314.885123] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e35b9f87-374c-4c12-a4ff-6e8abf3ee8c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.899646] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1314.899646] env[62235]: value = "task-1272356" [ 1314.899646] env[62235]: _type = "Task" [ 1314.899646] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.907244] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272356, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.409076] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272356, 'name': ReconfigVM_Task, 'duration_secs': 0.205146} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.409434] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-273541', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'name': 'volume-97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ba3567eb-9c7c-489a-bb85-eddef758b237', 'attached_at': '2024-10-12T06:31:01.000000', 'detached_at': '', 'volume_id': '97daf10b-3b9b-4cfc-a63d-44d617c3e098', 'serial': '97daf10b-3b9b-4cfc-a63d-44d617c3e098'} {{(pid=62235) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1315.409702] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1315.410444] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f011b6-4396-4d91-bf32-0be52499df8a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.416451] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1315.416667] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae6a6924-6247-40c8-a771-0af8e77881b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.480358] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1315.480588] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1315.480785] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleting the datastore file [datastore2] ba3567eb-9c7c-489a-bb85-eddef758b237 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1315.481085] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47bb8ac9-96c4-41b4-b20d-6257da63e435 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.487175] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1315.487175] env[62235]: value = "task-1272358" [ 1315.487175] env[62235]: _type = "Task" [ 1315.487175] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.494351] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.997336] env[62235]: DEBUG oslo_vmware.api [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272358, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.083737} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.997750] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1315.997804] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1315.997934] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1315.998120] env[62235]: INFO nova.compute.manager [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1315.998432] env[62235]: DEBUG oslo.service.loopingcall [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1315.998571] env[62235]: DEBUG nova.compute.manager [-] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1315.998665] env[62235]: DEBUG nova.network.neutron [-] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1316.416203] env[62235]: DEBUG nova.compute.manager [req-47b0e9e9-e96a-4ebd-9b14-42526f1123c8 req-048d0f18-1985-4bac-952f-ddbae43cc3bc service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Received event network-vif-deleted-773c38c4-89d2-4ca2-8d65-59200adf41ff {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1316.416445] env[62235]: INFO nova.compute.manager [req-47b0e9e9-e96a-4ebd-9b14-42526f1123c8 req-048d0f18-1985-4bac-952f-ddbae43cc3bc service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Neutron deleted interface 773c38c4-89d2-4ca2-8d65-59200adf41ff; detaching it from the instance and deleting it from the info cache [ 1316.416632] env[62235]: DEBUG nova.network.neutron [req-47b0e9e9-e96a-4ebd-9b14-42526f1123c8 req-048d0f18-1985-4bac-952f-ddbae43cc3bc service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.892433] env[62235]: DEBUG nova.network.neutron [-] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.919513] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c933e085-aeb0-448e-898d-1191b2b36f6b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.929101] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a854e2f-36e0-4869-8bb0-f721ede5e2a4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.953216] env[62235]: DEBUG nova.compute.manager [req-47b0e9e9-e96a-4ebd-9b14-42526f1123c8 req-048d0f18-1985-4bac-952f-ddbae43cc3bc service nova] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Detach interface failed, port_id=773c38c4-89d2-4ca2-8d65-59200adf41ff, reason: Instance ba3567eb-9c7c-489a-bb85-eddef758b237 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1317.395122] env[62235]: INFO nova.compute.manager [-] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Took 1.40 seconds to deallocate network for instance. [ 1317.938188] env[62235]: INFO nova.compute.manager [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Took 0.54 seconds to detach 1 volumes for instance. [ 1317.940514] env[62235]: DEBUG nova.compute.manager [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Deleting volume: 97daf10b-3b9b-4cfc-a63d-44d617c3e098 {{(pid=62235) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1318.480491] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1318.480893] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1318.480942] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1318.504808] env[62235]: INFO nova.scheduler.client.report [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted allocations for instance ba3567eb-9c7c-489a-bb85-eddef758b237 [ 1319.011910] env[62235]: DEBUG oslo_concurrency.lockutils [None req-c3eccb6f-8784-449f-acb6-c6525c76656d tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "ba3567eb-9c7c-489a-bb85-eddef758b237" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.243s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.331621] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "b7a22574-29f5-423b-b323-ea5eed8a65c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1319.331891] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1319.332131] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "b7a22574-29f5-423b-b323-ea5eed8a65c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1319.332324] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1319.332500] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.334729] env[62235]: INFO nova.compute.manager [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Terminating instance [ 1319.336442] env[62235]: DEBUG nova.compute.manager [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1319.336641] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1319.337524] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a42e0ea-a40e-492c-8bd3-0c99024243dd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.345226] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1319.345445] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73d6150c-35be-4fd4-8e4f-4cd4be9feb63 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.352058] env[62235]: DEBUG oslo_vmware.api [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1319.352058] env[62235]: value = "task-1272360" [ 1319.352058] env[62235]: _type = "Task" [ 1319.352058] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.359262] env[62235]: DEBUG oslo_vmware.api [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272360, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.862182] env[62235]: DEBUG oslo_vmware.api [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272360, 'name': PowerOffVM_Task, 'duration_secs': 0.197731} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.862458] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1319.862615] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1319.862854] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9f38ef9-24f1-438c-bcf1-f58be7e24554 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.921846] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1319.922094] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1319.922287] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleting the datastore file [datastore2] b7a22574-29f5-423b-b323-ea5eed8a65c4 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1319.922550] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6b0214e-f1a3-4578-8ad5-f48eba785b6f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.928380] env[62235]: DEBUG oslo_vmware.api [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1319.928380] env[62235]: value = "task-1272362" [ 1319.928380] env[62235]: _type = "Task" [ 1319.928380] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.935251] env[62235]: DEBUG oslo_vmware.api [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272362, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.438923] env[62235]: DEBUG oslo_vmware.api [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272362, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144163} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.439155] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1320.439352] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1320.439538] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1320.439719] env[62235]: INFO nova.compute.manager [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1320.439967] env[62235]: DEBUG oslo.service.loopingcall [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1320.440182] env[62235]: DEBUG nova.compute.manager [-] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1320.440279] env[62235]: DEBUG nova.network.neutron [-] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1320.659986] env[62235]: DEBUG nova.compute.manager [req-3d1173d5-6326-4662-a09e-a97cf3b2dece req-d5a67397-3c67-4d4d-9d4b-a653a4085665 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Received event network-vif-deleted-3cb0be04-88f5-48ea-bb89-6c93a10fb61b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1320.660185] env[62235]: INFO nova.compute.manager [req-3d1173d5-6326-4662-a09e-a97cf3b2dece req-d5a67397-3c67-4d4d-9d4b-a653a4085665 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Neutron deleted interface 3cb0be04-88f5-48ea-bb89-6c93a10fb61b; detaching it from the instance and deleting it from the info cache [ 1320.660367] env[62235]: DEBUG nova.network.neutron [req-3d1173d5-6326-4662-a09e-a97cf3b2dece req-d5a67397-3c67-4d4d-9d4b-a653a4085665 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1321.140032] env[62235]: DEBUG nova.network.neutron [-] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1321.162598] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44601f96-a1b6-42d3-84b8-3eeaeea6a75f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.174008] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49167212-5069-4cf1-8b71-d04981d3f7b1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.196939] env[62235]: DEBUG nova.compute.manager [req-3d1173d5-6326-4662-a09e-a97cf3b2dece req-d5a67397-3c67-4d4d-9d4b-a653a4085665 service nova] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Detach interface failed, port_id=3cb0be04-88f5-48ea-bb89-6c93a10fb61b, reason: Instance b7a22574-29f5-423b-b323-ea5eed8a65c4 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1321.641477] env[62235]: INFO nova.compute.manager [-] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Took 1.20 seconds to deallocate network for instance. [ 1322.147946] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1322.148369] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1322.148680] env[62235]: DEBUG nova.objects.instance [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'resources' on Instance uuid b7a22574-29f5-423b-b323-ea5eed8a65c4 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1322.669170] env[62235]: DEBUG nova.scheduler.client.report [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Refreshing inventories for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1322.683548] env[62235]: DEBUG nova.scheduler.client.report [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Updating ProviderTree inventory for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1322.683804] env[62235]: DEBUG nova.compute.provider_tree [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Updating inventory in ProviderTree for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1322.694203] env[62235]: DEBUG nova.scheduler.client.report [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Refreshing aggregate associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, aggregates: None {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1322.710984] env[62235]: DEBUG nova.scheduler.client.report [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Refreshing trait associations for resource provider 108a4390-ff68-4048-b61d-c7a9614ddc4f, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62235) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1322.746805] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98102bc2-ae83-4d53-b06b-53b4ec4b17fd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.754682] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8be6507-9d2c-4a51-b7bf-7004f6f503ce {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.783784] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a3e1604-3fec-4cb9-a894-36e82c4260fc {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.790354] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62af8476-34ac-49be-81a0-5974d801847e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.802707] env[62235]: DEBUG nova.compute.provider_tree [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1323.306235] env[62235]: DEBUG nova.scheduler.client.report [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1323.812236] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.664s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1323.830060] env[62235]: INFO nova.scheduler.client.report [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted allocations for instance b7a22574-29f5-423b-b323-ea5eed8a65c4 [ 1324.338626] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b038c02b-5182-4619-9956-9156545d3f6b tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "b7a22574-29f5-423b-b323-ea5eed8a65c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.007s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.773086] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1324.773370] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.279060] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.279060] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1326.284256] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Didn't find any instances for network info cache update. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1326.284543] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.284677] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.284821] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.284964] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.285117] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.285262] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.285396] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1326.285534] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.788494] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.788713] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.788855] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.789040] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1326.789971] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2c8a3a-c1fc-42c5-a53a-b3b3a9c33234 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.797886] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993a1a01-5d1a-4512-9930-f8b27b42d91a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.811557] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9892a18-f50b-4bea-b91f-4d611449ec5c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.817441] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9fbaf5-6532-449b-9b09-31f6d74dac87 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.524709] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180961MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1327.525097] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1327.525097] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.032116] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "0feb5395-f648-419a-998b-eff941c16464" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1328.032348] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "0feb5395-f648-419a-998b-eff941c16464" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.032540] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "0feb5395-f648-419a-998b-eff941c16464-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1328.032723] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "0feb5395-f648-419a-998b-eff941c16464-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.032895] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "0feb5395-f648-419a-998b-eff941c16464-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1328.034858] env[62235]: INFO nova.compute.manager [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Terminating instance [ 1328.036484] env[62235]: DEBUG nova.compute.manager [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1328.036681] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1328.037500] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c32b620-96fb-40da-a93f-46ae85dbd604 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.045224] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1328.045456] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d86da498-9816-494c-89cc-f7abc53f8fe2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.051791] env[62235]: DEBUG oslo_vmware.api [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1328.051791] env[62235]: value = "task-1272364" [ 1328.051791] env[62235]: _type = "Task" [ 1328.051791] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.059530] env[62235]: DEBUG oslo_vmware.api [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.552989] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 0feb5395-f648-419a-998b-eff941c16464 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1328.553252] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1328.553435] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1328.564199] env[62235]: DEBUG oslo_vmware.api [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272364, 'name': PowerOffVM_Task, 'duration_secs': 0.169868} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.564487] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1328.564658] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1328.564901] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5818a62a-8a12-4587-880b-afc7ffa460f6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.578360] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17bc84e-c309-48db-9ca3-531c402bf161 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.584888] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5cd37fa-661b-4f7b-b50c-d43296466f5f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.613414] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78d0dca-0bdb-423a-a6ca-eb42b13e8999 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.620404] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d19845-47ae-4730-ba62-695ec7e203d4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.634732] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1328.636727] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1328.636931] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1328.637124] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleting the datastore file [datastore1] 0feb5395-f648-419a-998b-eff941c16464 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1328.637567] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66ec3ed0-c7ba-4328-a819-d297354b0a47 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.643180] env[62235]: DEBUG oslo_vmware.api [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for the task: (returnval){ [ 1328.643180] env[62235]: value = "task-1272366" [ 1328.643180] env[62235]: _type = "Task" [ 1328.643180] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.651433] env[62235]: DEBUG oslo_vmware.api [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.140065] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1329.152898] env[62235]: DEBUG oslo_vmware.api [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Task: {'id': task-1272366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123513} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.153773] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1329.153966] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1329.154157] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1329.154354] env[62235]: INFO nova.compute.manager [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] [instance: 0feb5395-f648-419a-998b-eff941c16464] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1329.154600] env[62235]: DEBUG oslo.service.loopingcall [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1329.154794] env[62235]: DEBUG nova.compute.manager [-] [instance: 0feb5395-f648-419a-998b-eff941c16464] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1329.154886] env[62235]: DEBUG nova.network.neutron [-] [instance: 0feb5395-f648-419a-998b-eff941c16464] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1329.387062] env[62235]: DEBUG nova.compute.manager [req-129caa20-4f1b-445c-bfc6-d9f3da7c12bc req-9a257720-4206-4e3f-bb1d-fbba56e181f3 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Received event network-vif-deleted-8e87c7f6-11e9-49d5-8326-57e2140c176f {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1329.387186] env[62235]: INFO nova.compute.manager [req-129caa20-4f1b-445c-bfc6-d9f3da7c12bc req-9a257720-4206-4e3f-bb1d-fbba56e181f3 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Neutron deleted interface 8e87c7f6-11e9-49d5-8326-57e2140c176f; detaching it from the instance and deleting it from the info cache [ 1329.387416] env[62235]: DEBUG nova.network.neutron [req-129caa20-4f1b-445c-bfc6-d9f3da7c12bc req-9a257720-4206-4e3f-bb1d-fbba56e181f3 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.644664] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1329.645045] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.120s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1329.860460] env[62235]: DEBUG nova.network.neutron [-] [instance: 0feb5395-f648-419a-998b-eff941c16464] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.889590] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e38fd60-ee33-4a40-af4d-e467528717fa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.900928] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3ffd8e-c8e6-4922-a9ba-c2aacf528777 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.921289] env[62235]: DEBUG nova.compute.manager [req-129caa20-4f1b-445c-bfc6-d9f3da7c12bc req-9a257720-4206-4e3f-bb1d-fbba56e181f3 service nova] [instance: 0feb5395-f648-419a-998b-eff941c16464] Detach interface failed, port_id=8e87c7f6-11e9-49d5-8326-57e2140c176f, reason: Instance 0feb5395-f648-419a-998b-eff941c16464 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1330.363843] env[62235]: INFO nova.compute.manager [-] [instance: 0feb5395-f648-419a-998b-eff941c16464] Took 1.21 seconds to deallocate network for instance. [ 1330.870508] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.870799] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.871027] env[62235]: DEBUG nova.objects.instance [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lazy-loading 'resources' on Instance uuid 0feb5395-f648-419a-998b-eff941c16464 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1331.406549] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a82d417-48f0-4828-a600-ffecc3473384 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.414986] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c719ece-81fe-4ac0-9ba2-dc1ab06264af {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.443438] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3484ec29-efb4-4895-b0e2-155d1d92fde7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.450167] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832c2303-513c-4c45-8fad-b9e156ecf644 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.462564] env[62235]: DEBUG nova.compute.provider_tree [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1331.965952] env[62235]: DEBUG nova.scheduler.client.report [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1332.470556] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.489980] env[62235]: INFO nova.scheduler.client.report [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Deleted allocations for instance 0feb5395-f648-419a-998b-eff941c16464 [ 1332.997534] env[62235]: DEBUG oslo_concurrency.lockutils [None req-1d1142b6-e6f0-4678-b686-0e813483d2d0 tempest-ServerActionsTestOtherA-28038047 tempest-ServerActionsTestOtherA-28038047-project-member] Lock "0feb5395-f648-419a-998b-eff941c16464" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.965s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1337.959582] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8e177c90-8d68-4d70-9134-d2635abdfe7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.959919] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1338.463116] env[62235]: DEBUG nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1338.984494] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.984791] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1338.986209] env[62235]: INFO nova.compute.claims [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1340.020801] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d9c69f-6294-4627-a4e8-c69272dac178 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.028562] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e58de84-464e-428d-8e5d-a4b9011613c4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.057510] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295a025d-ee02-476b-8f99-8816ce1c0312 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.064179] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184cb53e-0f9d-40df-afac-f068e7c62780 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.077690] env[62235]: DEBUG nova.compute.provider_tree [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1340.580823] env[62235]: DEBUG nova.scheduler.client.report [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1341.086184] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.101s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1341.086734] env[62235]: DEBUG nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1341.591504] env[62235]: DEBUG nova.compute.utils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1341.593012] env[62235]: DEBUG nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1341.593273] env[62235]: DEBUG nova.network.neutron [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1341.629742] env[62235]: DEBUG nova.policy [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8de8d8815704b1ca99fe1f48d746109', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd439d7dfff414de6ba781d2a7d464120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1341.854293] env[62235]: DEBUG nova.network.neutron [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Successfully created port: 909b7c34-704b-47db-a0c2-a33b6e25a8da {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1342.096451] env[62235]: DEBUG nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1343.105102] env[62235]: DEBUG nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1343.132465] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1343.132749] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1343.132960] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1343.133191] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1343.133348] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1343.133549] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1343.133780] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1343.133945] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1343.134127] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1343.134293] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1343.134475] env[62235]: DEBUG nova.virt.hardware [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1343.135357] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8728c708-8ad9-4e84-94f4-920d1aefccd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.143098] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f644ef-a706-4108-a334-d40f005838a3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.208672] env[62235]: DEBUG nova.compute.manager [req-86260d29-0df7-4617-97c4-db702a67ec05 req-884049cc-324e-4315-887b-268d193d4094 service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Received event network-vif-plugged-909b7c34-704b-47db-a0c2-a33b6e25a8da {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1343.208896] env[62235]: DEBUG oslo_concurrency.lockutils [req-86260d29-0df7-4617-97c4-db702a67ec05 req-884049cc-324e-4315-887b-268d193d4094 service nova] Acquiring lock "8e177c90-8d68-4d70-9134-d2635abdfe7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1343.209174] env[62235]: DEBUG oslo_concurrency.lockutils [req-86260d29-0df7-4617-97c4-db702a67ec05 req-884049cc-324e-4315-887b-268d193d4094 service nova] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1343.209291] env[62235]: DEBUG oslo_concurrency.lockutils [req-86260d29-0df7-4617-97c4-db702a67ec05 req-884049cc-324e-4315-887b-268d193d4094 service nova] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1343.209430] env[62235]: DEBUG nova.compute.manager [req-86260d29-0df7-4617-97c4-db702a67ec05 req-884049cc-324e-4315-887b-268d193d4094 service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] No waiting events found dispatching network-vif-plugged-909b7c34-704b-47db-a0c2-a33b6e25a8da {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1343.209597] env[62235]: WARNING nova.compute.manager [req-86260d29-0df7-4617-97c4-db702a67ec05 req-884049cc-324e-4315-887b-268d193d4094 service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Received unexpected event network-vif-plugged-909b7c34-704b-47db-a0c2-a33b6e25a8da for instance with vm_state building and task_state spawning. [ 1343.290965] env[62235]: DEBUG nova.network.neutron [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Successfully updated port: 909b7c34-704b-47db-a0c2-a33b6e25a8da {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1343.794083] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1343.794264] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1343.794381] env[62235]: DEBUG nova.network.neutron [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1344.325370] env[62235]: DEBUG nova.network.neutron [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1344.443698] env[62235]: DEBUG nova.network.neutron [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updating instance_info_cache with network_info: [{"id": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "address": "fa:16:3e:73:ba:c1", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap909b7c34-70", "ovs_interfaceid": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.946399] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1344.946746] env[62235]: DEBUG nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Instance network_info: |[{"id": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "address": "fa:16:3e:73:ba:c1", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap909b7c34-70", "ovs_interfaceid": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1344.947203] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:ba:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '849fc06e-dfc2-470f-8490-034590682ea7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '909b7c34-704b-47db-a0c2-a33b6e25a8da', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1344.954510] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating folder: Project (d439d7dfff414de6ba781d2a7d464120). Parent ref: group-v273362. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1344.954794] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99f335c7-762a-4ec5-aa11-6b440f854576 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.967123] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created folder: Project (d439d7dfff414de6ba781d2a7d464120) in parent group-v273362. [ 1344.967323] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating folder: Instances. Parent ref: group-v273543. {{(pid=62235) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1344.967879] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c41b12c4-803e-4de0-9a63-5825c0da15f4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.976555] env[62235]: INFO nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created folder: Instances in parent group-v273543. [ 1344.976780] env[62235]: DEBUG oslo.service.loopingcall [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1344.976966] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1344.977178] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-902b5368-28d9-49cf-b817-6c5f472bb53d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.995164] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1344.995164] env[62235]: value = "task-1272369" [ 1344.995164] env[62235]: _type = "Task" [ 1344.995164] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.001997] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272369, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.236725] env[62235]: DEBUG nova.compute.manager [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Received event network-changed-909b7c34-704b-47db-a0c2-a33b6e25a8da {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1345.236971] env[62235]: DEBUG nova.compute.manager [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Refreshing instance network info cache due to event network-changed-909b7c34-704b-47db-a0c2-a33b6e25a8da. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1345.237210] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] Acquiring lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1345.237376] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] Acquired lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1345.237559] env[62235]: DEBUG nova.network.neutron [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Refreshing network info cache for port 909b7c34-704b-47db-a0c2-a33b6e25a8da {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1345.505711] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272369, 'name': CreateVM_Task, 'duration_secs': 0.285125} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.506080] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1345.506532] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1345.506703] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1345.507047] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1345.507290] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a07b6f16-2b63-4707-9ce6-ae6ef07b076c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.511638] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1345.511638] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525871cc-fcfb-0835-e6c9-807fe6a6c33d" [ 1345.511638] env[62235]: _type = "Task" [ 1345.511638] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.518616] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525871cc-fcfb-0835-e6c9-807fe6a6c33d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.911576] env[62235]: DEBUG nova.network.neutron [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updated VIF entry in instance network info cache for port 909b7c34-704b-47db-a0c2-a33b6e25a8da. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1345.911923] env[62235]: DEBUG nova.network.neutron [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updating instance_info_cache with network_info: [{"id": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "address": "fa:16:3e:73:ba:c1", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap909b7c34-70", "ovs_interfaceid": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1346.021690] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]525871cc-fcfb-0835-e6c9-807fe6a6c33d, 'name': SearchDatastore_Task, 'duration_secs': 0.009984} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.021974] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1346.022230] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1346.022468] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1346.022619] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1346.022801] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1346.023076] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7016fbb-530a-4f7e-9ecf-b25479c98e2c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.031022] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1346.031202] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1346.031915] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3501f773-ceff-477b-b30b-86c5fdc392da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.036885] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1346.036885] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5295f7c2-3c4f-844e-3de7-b174b305c112" [ 1346.036885] env[62235]: _type = "Task" [ 1346.036885] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.043693] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5295f7c2-3c4f-844e-3de7-b174b305c112, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.414534] env[62235]: DEBUG oslo_concurrency.lockutils [req-5fa5669a-cec7-4560-99fe-35ccc6f6ae7c req-2b558522-8aad-4a3b-a865-81033ff8fa5b service nova] Releasing lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1346.546611] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5295f7c2-3c4f-844e-3de7-b174b305c112, 'name': SearchDatastore_Task, 'duration_secs': 0.007958} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.547341] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-244e1ad5-f559-4eea-bbdd-c1066b12b116 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.552208] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1346.552208] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522f50f2-76b1-dbc4-13f1-75dea15cbbc9" [ 1346.552208] env[62235]: _type = "Task" [ 1346.552208] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.558979] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522f50f2-76b1-dbc4-13f1-75dea15cbbc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.062821] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]522f50f2-76b1-dbc4-13f1-75dea15cbbc9, 'name': SearchDatastore_Task, 'duration_secs': 0.009358} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.063106] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1347.063371] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 8e177c90-8d68-4d70-9134-d2635abdfe7f/8e177c90-8d68-4d70-9134-d2635abdfe7f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1347.063657] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34b60f75-b390-479b-8928-b30cdf5fc3b9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.070153] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1347.070153] env[62235]: value = "task-1272370" [ 1347.070153] env[62235]: _type = "Task" [ 1347.070153] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.077711] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.579681] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272370, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421192} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.580130] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 8e177c90-8d68-4d70-9134-d2635abdfe7f/8e177c90-8d68-4d70-9134-d2635abdfe7f.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1347.580235] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1347.580461] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a8cee29-a91c-4c17-97ce-9a10ca5539e2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.587560] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1347.587560] env[62235]: value = "task-1272371" [ 1347.587560] env[62235]: _type = "Task" [ 1347.587560] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.594491] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.097853] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058752} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.099073] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1348.099188] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc25236a-33ba-4458-a6a7-643f873ff4e0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.120581] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 8e177c90-8d68-4d70-9134-d2635abdfe7f/8e177c90-8d68-4d70-9134-d2635abdfe7f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1348.120798] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-633c68a6-4994-40dd-880e-dbed7f11909d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.140054] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1348.140054] env[62235]: value = "task-1272372" [ 1348.140054] env[62235]: _type = "Task" [ 1348.140054] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.147265] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.649909] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272372, 'name': ReconfigVM_Task, 'duration_secs': 0.278938} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.650226] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 8e177c90-8d68-4d70-9134-d2635abdfe7f/8e177c90-8d68-4d70-9134-d2635abdfe7f.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1348.650836] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ed1cfc5-1839-463d-b521-4954fd279aaf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.657016] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1348.657016] env[62235]: value = "task-1272373" [ 1348.657016] env[62235]: _type = "Task" [ 1348.657016] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.664347] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272373, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.166933] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272373, 'name': Rename_Task, 'duration_secs': 0.132807} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.167228] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1349.167501] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-794d17b7-6914-4d74-847d-4f7136f26be8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.174230] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1349.174230] env[62235]: value = "task-1272374" [ 1349.174230] env[62235]: _type = "Task" [ 1349.174230] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.181295] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.684015] env[62235]: DEBUG oslo_vmware.api [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272374, 'name': PowerOnVM_Task, 'duration_secs': 0.430083} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.684397] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1349.684503] env[62235]: INFO nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Took 6.58 seconds to spawn the instance on the hypervisor. [ 1349.684682] env[62235]: DEBUG nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1349.685421] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b044742-f4b8-4c38-ae1d-7b9670945741 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.203387] env[62235]: INFO nova.compute.manager [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Took 11.24 seconds to build instance. [ 1350.706208] env[62235]: DEBUG oslo_concurrency.lockutils [None req-47d29681-6d69-441c-800e-93203fbbddca tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.746s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1351.862829] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1351.863128] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.365775] env[62235]: DEBUG nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1352.887782] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.888117] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.889628] env[62235]: INFO nova.compute.claims [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1353.934551] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b652059-07b7-4746-ba98-f645d124bb30 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.942139] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d932341b-c0a3-484f-b963-36ed158fd820 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.971930] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd541e7-d0d3-492b-a752-75639965ad2a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.978974] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4446a77f-5deb-42b0-a252-4c9276f69f52 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.991916] env[62235]: DEBUG nova.compute.provider_tree [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1354.494977] env[62235]: DEBUG nova.scheduler.client.report [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1355.000167] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.112s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1355.000724] env[62235]: DEBUG nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1355.506209] env[62235]: DEBUG nova.compute.utils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1355.507657] env[62235]: DEBUG nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1355.508029] env[62235]: DEBUG nova.network.neutron [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1355.562915] env[62235]: DEBUG nova.policy [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8de8d8815704b1ca99fe1f48d746109', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd439d7dfff414de6ba781d2a7d464120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1355.803678] env[62235]: DEBUG nova.network.neutron [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Successfully created port: e1ef049f-546a-480e-9d89-e2a68fcca0aa {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1356.011741] env[62235]: DEBUG nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1357.021249] env[62235]: DEBUG nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1357.046243] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1357.046533] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1357.046700] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1357.046890] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1357.047101] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1357.047232] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1357.047446] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1357.047610] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1357.047780] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1357.048320] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1357.048320] env[62235]: DEBUG nova.virt.hardware [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1357.048976] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e56db53-2105-4c4a-b352-ec4fbd7a572a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.058057] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4af7eb-ea52-4de0-8d52-fcf1da23891f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.165572] env[62235]: DEBUG nova.compute.manager [req-0fe6d01a-9074-4ebb-980a-ae53daab5c77 req-f3301f70-f8f1-4888-8c46-57cf457a09b9 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Received event network-vif-plugged-e1ef049f-546a-480e-9d89-e2a68fcca0aa {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1357.166138] env[62235]: DEBUG oslo_concurrency.lockutils [req-0fe6d01a-9074-4ebb-980a-ae53daab5c77 req-f3301f70-f8f1-4888-8c46-57cf457a09b9 service nova] Acquiring lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1357.166627] env[62235]: DEBUG oslo_concurrency.lockutils [req-0fe6d01a-9074-4ebb-980a-ae53daab5c77 req-f3301f70-f8f1-4888-8c46-57cf457a09b9 service nova] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.167086] env[62235]: DEBUG oslo_concurrency.lockutils [req-0fe6d01a-9074-4ebb-980a-ae53daab5c77 req-f3301f70-f8f1-4888-8c46-57cf457a09b9 service nova] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.167539] env[62235]: DEBUG nova.compute.manager [req-0fe6d01a-9074-4ebb-980a-ae53daab5c77 req-f3301f70-f8f1-4888-8c46-57cf457a09b9 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] No waiting events found dispatching network-vif-plugged-e1ef049f-546a-480e-9d89-e2a68fcca0aa {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1357.167970] env[62235]: WARNING nova.compute.manager [req-0fe6d01a-9074-4ebb-980a-ae53daab5c77 req-f3301f70-f8f1-4888-8c46-57cf457a09b9 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Received unexpected event network-vif-plugged-e1ef049f-546a-480e-9d89-e2a68fcca0aa for instance with vm_state building and task_state spawning. [ 1357.246971] env[62235]: DEBUG nova.network.neutron [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Successfully updated port: e1ef049f-546a-480e-9d89-e2a68fcca0aa {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1357.751039] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "refresh_cache-8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.751225] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "refresh_cache-8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.751360] env[62235]: DEBUG nova.network.neutron [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1358.283028] env[62235]: DEBUG nova.network.neutron [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1358.403956] env[62235]: DEBUG nova.network.neutron [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Updating instance_info_cache with network_info: [{"id": "e1ef049f-546a-480e-9d89-e2a68fcca0aa", "address": "fa:16:3e:26:f5:6e", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1ef049f-54", "ovs_interfaceid": "e1ef049f-546a-480e-9d89-e2a68fcca0aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.906728] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "refresh_cache-8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1358.907115] env[62235]: DEBUG nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Instance network_info: |[{"id": "e1ef049f-546a-480e-9d89-e2a68fcca0aa", "address": "fa:16:3e:26:f5:6e", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1ef049f-54", "ovs_interfaceid": "e1ef049f-546a-480e-9d89-e2a68fcca0aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1358.907564] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:f5:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '849fc06e-dfc2-470f-8490-034590682ea7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1ef049f-546a-480e-9d89-e2a68fcca0aa', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1358.914929] env[62235]: DEBUG oslo.service.loopingcall [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1358.915158] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1358.915738] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-521b82f2-76af-4853-a5bd-068b535eed37 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.936017] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1358.936017] env[62235]: value = "task-1272375" [ 1358.936017] env[62235]: _type = "Task" [ 1358.936017] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.943604] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272375, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.191885] env[62235]: DEBUG nova.compute.manager [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Received event network-changed-e1ef049f-546a-480e-9d89-e2a68fcca0aa {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1359.192122] env[62235]: DEBUG nova.compute.manager [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Refreshing instance network info cache due to event network-changed-e1ef049f-546a-480e-9d89-e2a68fcca0aa. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1359.192348] env[62235]: DEBUG oslo_concurrency.lockutils [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] Acquiring lock "refresh_cache-8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1359.192496] env[62235]: DEBUG oslo_concurrency.lockutils [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] Acquired lock "refresh_cache-8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.192655] env[62235]: DEBUG nova.network.neutron [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Refreshing network info cache for port e1ef049f-546a-480e-9d89-e2a68fcca0aa {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1359.445963] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272375, 'name': CreateVM_Task, 'duration_secs': 0.284144} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.446350] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1359.446777] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1359.446950] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.447299] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1359.447556] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5194b131-24ea-45be-a8a6-5b19daab3e4d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.452371] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1359.452371] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5278990d-259f-3cbd-218e-24702f34ec1e" [ 1359.452371] env[62235]: _type = "Task" [ 1359.452371] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.460466] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5278990d-259f-3cbd-218e-24702f34ec1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.890203] env[62235]: DEBUG nova.network.neutron [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Updated VIF entry in instance network info cache for port e1ef049f-546a-480e-9d89-e2a68fcca0aa. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1359.890600] env[62235]: DEBUG nova.network.neutron [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Updating instance_info_cache with network_info: [{"id": "e1ef049f-546a-480e-9d89-e2a68fcca0aa", "address": "fa:16:3e:26:f5:6e", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1ef049f-54", "ovs_interfaceid": "e1ef049f-546a-480e-9d89-e2a68fcca0aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1359.963246] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5278990d-259f-3cbd-218e-24702f34ec1e, 'name': SearchDatastore_Task, 'duration_secs': 0.009321} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.963573] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1359.963850] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1359.964110] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1359.964266] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.964449] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1359.964717] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a88e666-d163-4c60-8b0b-e02cba8088e5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.973482] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1359.973658] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1359.974382] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf8b9243-541b-4f5f-9344-2c57af9c7693 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.979705] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1359.979705] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c9ecfd-fd3f-53db-9dfc-ecf898c3b9da" [ 1359.979705] env[62235]: _type = "Task" [ 1359.979705] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.987427] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c9ecfd-fd3f-53db-9dfc-ecf898c3b9da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.393511] env[62235]: DEBUG oslo_concurrency.lockutils [req-dbcc0e7c-e754-48e0-a5e5-d7b9377a5d16 req-a9bf7d0e-c472-4605-9e1e-29203a0c1223 service nova] Releasing lock "refresh_cache-8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1360.491409] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52c9ecfd-fd3f-53db-9dfc-ecf898c3b9da, 'name': SearchDatastore_Task, 'duration_secs': 0.008981} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.492267] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1a29706-7989-4f19-9a90-b3ede31c8847 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.498259] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1360.498259] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5216fa10-1c96-0f0b-8905-1b74406ac631" [ 1360.498259] env[62235]: _type = "Task" [ 1360.498259] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.506332] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5216fa10-1c96-0f0b-8905-1b74406ac631, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.008931] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5216fa10-1c96-0f0b-8905-1b74406ac631, 'name': SearchDatastore_Task, 'duration_secs': 0.010475} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.009223] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1361.009486] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b/8c4ed32c-47a9-4a2b-b8af-f4371c229c3b.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1361.009743] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-188f0478-9da5-4bc9-8474-1986f056942f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.016141] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1361.016141] env[62235]: value = "task-1272376" [ 1361.016141] env[62235]: _type = "Task" [ 1361.016141] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.023377] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272376, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.525442] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272376, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411532} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.525788] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b/8c4ed32c-47a9-4a2b-b8af-f4371c229c3b.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1361.526113] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1361.526420] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a569d791-8ee1-4450-8936-61921e8f3355 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.533314] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1361.533314] env[62235]: value = "task-1272377" [ 1361.533314] env[62235]: _type = "Task" [ 1361.533314] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.540418] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.042437] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272377, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063754} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.042780] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1362.043549] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faacd10-0926-4579-9e66-4fbf84e2dfe5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.064503] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b/8c4ed32c-47a9-4a2b-b8af-f4371c229c3b.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1362.064774] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab2d8f53-d28d-43db-b370-b97ee3eef530 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.084167] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1362.084167] env[62235]: value = "task-1272378" [ 1362.084167] env[62235]: _type = "Task" [ 1362.084167] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.091093] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272378, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.594167] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272378, 'name': ReconfigVM_Task, 'duration_secs': 0.316132} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.594457] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b/8c4ed32c-47a9-4a2b-b8af-f4371c229c3b.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1362.595098] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da39886d-6670-4336-b0f9-9f78107fabb6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.601960] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1362.601960] env[62235]: value = "task-1272379" [ 1362.601960] env[62235]: _type = "Task" [ 1362.601960] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.610131] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272379, 'name': Rename_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.111490] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272379, 'name': Rename_Task, 'duration_secs': 0.125802} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.111780] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1363.112038] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38cd384a-2046-411e-974d-daa8a43f7d81 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.117660] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1363.117660] env[62235]: value = "task-1272380" [ 1363.117660] env[62235]: _type = "Task" [ 1363.117660] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.124830] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.627678] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272380, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.127628] env[62235]: DEBUG oslo_vmware.api [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272380, 'name': PowerOnVM_Task, 'duration_secs': 0.603962} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.127903] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1364.128159] env[62235]: INFO nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Took 7.11 seconds to spawn the instance on the hypervisor. [ 1364.128350] env[62235]: DEBUG nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1364.129094] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad30dcd-3f2a-4842-9aff-bc4f5d1282d2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.647149] env[62235]: INFO nova.compute.manager [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Took 11.78 seconds to build instance. [ 1365.144499] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1365.149070] env[62235]: DEBUG oslo_concurrency.lockutils [None req-acd3aaed-3826-4b79-8bdb-01706b603edf tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.286s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1365.149315] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.005s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1365.149526] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1365.149723] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1365.149891] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1365.152078] env[62235]: INFO nova.compute.manager [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Terminating instance [ 1365.153931] env[62235]: DEBUG nova.compute.manager [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1365.154154] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1365.154978] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed7c648-aa60-4c67-ad5f-654eab86eb96 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.162630] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1365.162856] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec228e72-ecdc-4599-8574-49983fbaeb80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.169467] env[62235]: DEBUG oslo_vmware.api [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1365.169467] env[62235]: value = "task-1272381" [ 1365.169467] env[62235]: _type = "Task" [ 1365.169467] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.177930] env[62235]: DEBUG oslo_vmware.api [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.679735] env[62235]: DEBUG oslo_vmware.api [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272381, 'name': PowerOffVM_Task, 'duration_secs': 0.166843} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.680116] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1365.680186] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1365.680426] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d7f11c7-f2f7-452e-b04d-3e048161f41b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.738949] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1365.739146] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1365.739310] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleting the datastore file [datastore1] 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1365.739570] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3bd3162-4f7a-4f80-a17f-5d92dd7fc25f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.746008] env[62235]: DEBUG oslo_vmware.api [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1365.746008] env[62235]: value = "task-1272383" [ 1365.746008] env[62235]: _type = "Task" [ 1365.746008] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.753366] env[62235]: DEBUG oslo_vmware.api [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.256374] env[62235]: DEBUG oslo_vmware.api [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272383, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127306} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.256640] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1366.256827] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1366.257017] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1366.257206] env[62235]: INFO nova.compute.manager [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1366.257446] env[62235]: DEBUG oslo.service.loopingcall [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1366.257643] env[62235]: DEBUG nova.compute.manager [-] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1366.257734] env[62235]: DEBUG nova.network.neutron [-] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1366.485298] env[62235]: DEBUG nova.compute.manager [req-456d4e0c-f61a-43ff-9c55-3e0dc6bcbb83 req-8a7dd80d-87db-4a3e-aaf3-8a68369b8dec service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Received event network-vif-deleted-e1ef049f-546a-480e-9d89-e2a68fcca0aa {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1366.485550] env[62235]: INFO nova.compute.manager [req-456d4e0c-f61a-43ff-9c55-3e0dc6bcbb83 req-8a7dd80d-87db-4a3e-aaf3-8a68369b8dec service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Neutron deleted interface e1ef049f-546a-480e-9d89-e2a68fcca0aa; detaching it from the instance and deleting it from the info cache [ 1366.485763] env[62235]: DEBUG nova.network.neutron [req-456d4e0c-f61a-43ff-9c55-3e0dc6bcbb83 req-8a7dd80d-87db-4a3e-aaf3-8a68369b8dec service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1366.963961] env[62235]: DEBUG nova.network.neutron [-] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1366.988413] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5648d55b-961b-446e-9d92-a4b355e5d70e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.998011] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac3616a-5211-4c35-bb47-1858869653a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.021932] env[62235]: DEBUG nova.compute.manager [req-456d4e0c-f61a-43ff-9c55-3e0dc6bcbb83 req-8a7dd80d-87db-4a3e-aaf3-8a68369b8dec service nova] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Detach interface failed, port_id=e1ef049f-546a-480e-9d89-e2a68fcca0aa, reason: Instance 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1367.466367] env[62235]: INFO nova.compute.manager [-] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Took 1.21 seconds to deallocate network for instance. [ 1367.972518] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1367.972801] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1367.973028] env[62235]: DEBUG nova.objects.instance [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lazy-loading 'resources' on Instance uuid 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1368.515715] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d93e105-d4e4-4fdf-8355-e8b7a353f679 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.523180] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28631038-d1dc-4d61-8ee3-d49613573879 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.553393] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dbf9c6e-f2f3-4b06-9958-5642d688870a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.559891] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae2cbac-be01-4ffa-b3db-1edcbd944534 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.572504] env[62235]: DEBUG nova.compute.provider_tree [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1369.075346] env[62235]: DEBUG nova.scheduler.client.report [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1369.580998] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1369.600482] env[62235]: INFO nova.scheduler.client.report [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted allocations for instance 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b [ 1370.108348] env[62235]: DEBUG oslo_concurrency.lockutils [None req-2b0bbbe9-8e3d-4550-b09b-96210fb14d25 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8c4ed32c-47a9-4a2b-b8af-f4371c229c3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.959s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1371.215076] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1371.215403] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1371.718105] env[62235]: DEBUG nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1372.240768] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.241060] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.242598] env[62235]: INFO nova.compute.claims [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1372.321737] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.285534] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f812cfe2-e25e-45c4-9596-df5c3b6ba7eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.292962] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b2ce96-3266-4a5a-ab63-6ae8562345d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.322346] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e06831-21d5-4d8f-a36a-dacb37cad334 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.329143] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a647092-f522-4e52-8fcf-b12bb2a6ee42 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.341529] env[62235]: DEBUG nova.compute.provider_tree [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1373.844567] env[62235]: DEBUG nova.scheduler.client.report [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1374.349979] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.109s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.350544] env[62235]: DEBUG nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1374.824201] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.824372] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Cleaning up deleted instances {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1374.855643] env[62235]: DEBUG nova.compute.utils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1374.857036] env[62235]: DEBUG nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1374.857218] env[62235]: DEBUG nova.network.neutron [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1374.904155] env[62235]: DEBUG nova.policy [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8de8d8815704b1ca99fe1f48d746109', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd439d7dfff414de6ba781d2a7d464120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1375.136714] env[62235]: DEBUG nova.network.neutron [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Successfully created port: 7d09f3b7-e012-44b1-81ec-5ba677f6848e {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1375.330104] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] There are 13 instances to clean {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1375.330610] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8c4ed32c-47a9-4a2b-b8af-f4371c229c3b] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1375.360348] env[62235]: DEBUG nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1375.833702] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ba3567eb-9c7c-489a-bb85-eddef758b237] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1376.336974] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b7a22574-29f5-423b-b323-ea5eed8a65c4] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1376.369582] env[62235]: DEBUG nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1376.393965] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1376.394303] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1376.394472] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1376.394661] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1376.394804] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1376.394981] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1376.395172] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1376.395336] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1376.395503] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1376.395667] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1376.395837] env[62235]: DEBUG nova.virt.hardware [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1376.396726] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473c569e-bce1-46ba-959e-c81bf84171f5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.404826] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d6e139-3520-4052-b845-5f1575690384 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.508580] env[62235]: DEBUG nova.compute.manager [req-b1c864bb-fb5e-4b21-bee8-e093464cd4cf req-3dd59e31-5697-43f4-a3d2-2d9dadf6be12 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Received event network-vif-plugged-7d09f3b7-e012-44b1-81ec-5ba677f6848e {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1376.508812] env[62235]: DEBUG oslo_concurrency.lockutils [req-b1c864bb-fb5e-4b21-bee8-e093464cd4cf req-3dd59e31-5697-43f4-a3d2-2d9dadf6be12 service nova] Acquiring lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1376.509035] env[62235]: DEBUG oslo_concurrency.lockutils [req-b1c864bb-fb5e-4b21-bee8-e093464cd4cf req-3dd59e31-5697-43f4-a3d2-2d9dadf6be12 service nova] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1376.509213] env[62235]: DEBUG oslo_concurrency.lockutils [req-b1c864bb-fb5e-4b21-bee8-e093464cd4cf req-3dd59e31-5697-43f4-a3d2-2d9dadf6be12 service nova] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1376.509388] env[62235]: DEBUG nova.compute.manager [req-b1c864bb-fb5e-4b21-bee8-e093464cd4cf req-3dd59e31-5697-43f4-a3d2-2d9dadf6be12 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] No waiting events found dispatching network-vif-plugged-7d09f3b7-e012-44b1-81ec-5ba677f6848e {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1376.509555] env[62235]: WARNING nova.compute.manager [req-b1c864bb-fb5e-4b21-bee8-e093464cd4cf req-3dd59e31-5697-43f4-a3d2-2d9dadf6be12 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Received unexpected event network-vif-plugged-7d09f3b7-e012-44b1-81ec-5ba677f6848e for instance with vm_state building and task_state spawning. [ 1376.586047] env[62235]: DEBUG nova.network.neutron [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Successfully updated port: 7d09f3b7-e012-44b1-81ec-5ba677f6848e {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1376.840593] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: b39057fd-7acc-437a-bb61-4b84ba4df435] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.089020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "refresh_cache-ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.089020] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "refresh_cache-ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.089118] env[62235]: DEBUG nova.network.neutron [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1377.343763] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: a293951c-183d-4bfe-b41e-2b3408f2eee3] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.619864] env[62235]: DEBUG nova.network.neutron [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1377.740991] env[62235]: DEBUG nova.network.neutron [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Updating instance_info_cache with network_info: [{"id": "7d09f3b7-e012-44b1-81ec-5ba677f6848e", "address": "fa:16:3e:f1:41:64", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d09f3b7-e0", "ovs_interfaceid": "7d09f3b7-e012-44b1-81ec-5ba677f6848e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1377.846679] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: f7ef033e-ab43-4b89-b331-cb234a373991] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.244436] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "refresh_cache-ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1378.244769] env[62235]: DEBUG nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Instance network_info: |[{"id": "7d09f3b7-e012-44b1-81ec-5ba677f6848e", "address": "fa:16:3e:f1:41:64", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d09f3b7-e0", "ovs_interfaceid": "7d09f3b7-e012-44b1-81ec-5ba677f6848e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1378.245285] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:41:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '849fc06e-dfc2-470f-8490-034590682ea7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d09f3b7-e012-44b1-81ec-5ba677f6848e', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1378.254537] env[62235]: DEBUG oslo.service.loopingcall [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1378.254754] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1378.254976] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b806ee3f-f588-4c45-9dc4-006dd77b18ca {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.275790] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1378.275790] env[62235]: value = "task-1272384" [ 1378.275790] env[62235]: _type = "Task" [ 1378.275790] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.283214] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272384, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.350000] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: e6421e2f-17cf-4312-a122-067cd032c066] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.535484] env[62235]: DEBUG nova.compute.manager [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Received event network-changed-7d09f3b7-e012-44b1-81ec-5ba677f6848e {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1378.535697] env[62235]: DEBUG nova.compute.manager [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Refreshing instance network info cache due to event network-changed-7d09f3b7-e012-44b1-81ec-5ba677f6848e. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1378.537609] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] Acquiring lock "refresh_cache-ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1378.537816] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] Acquired lock "refresh_cache-ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1378.537996] env[62235]: DEBUG nova.network.neutron [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Refreshing network info cache for port 7d09f3b7-e012-44b1-81ec-5ba677f6848e {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1378.785249] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272384, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.854146] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: f1ef38cf-4f4b-4c53-a69d-5c03493a411f] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.227759] env[62235]: DEBUG nova.network.neutron [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Updated VIF entry in instance network info cache for port 7d09f3b7-e012-44b1-81ec-5ba677f6848e. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1379.228186] env[62235]: DEBUG nova.network.neutron [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Updating instance_info_cache with network_info: [{"id": "7d09f3b7-e012-44b1-81ec-5ba677f6848e", "address": "fa:16:3e:f1:41:64", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d09f3b7-e0", "ovs_interfaceid": "7d09f3b7-e012-44b1-81ec-5ba677f6848e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1379.285844] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272384, 'name': CreateVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.357182] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 5e52ce21-04a1-41db-b62e-2932bf91538e] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.730986] env[62235]: DEBUG oslo_concurrency.lockutils [req-cc3d08e8-16bd-4edb-9be0-7819b233a417 req-f7e7607d-6fad-4e4c-b74b-a4624ef60485 service nova] Releasing lock "refresh_cache-ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1379.786611] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272384, 'name': CreateVM_Task, 'duration_secs': 1.288063} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.786952] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1379.793565] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1379.793736] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1379.794074] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1379.794345] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98db3b98-5eb7-418f-aee7-d321e4feb725 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.798388] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1379.798388] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ee38d-2db4-b407-a034-6bfe25c88af1" [ 1379.798388] env[62235]: _type = "Task" [ 1379.798388] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.805419] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ee38d-2db4-b407-a034-6bfe25c88af1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.860046] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: a993bed1-e346-47e7-9f7f-abdfb6685196] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.309339] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]529ee38d-2db4-b407-a034-6bfe25c88af1, 'name': SearchDatastore_Task, 'duration_secs': 0.009866} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.309639] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1380.309879] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1380.310125] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1380.310280] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1380.310463] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1380.310723] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1b78d0d-1f7b-4eaf-85b0-19c9e24696d3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.318274] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1380.318449] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1380.319121] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-552bc969-cec0-4099-8adf-7d178b0b140b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.323526] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1380.323526] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527bc6ce-32a0-b6fd-f8a9-e9b200fa8447" [ 1380.323526] env[62235]: _type = "Task" [ 1380.323526] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.330402] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527bc6ce-32a0-b6fd-f8a9-e9b200fa8447, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.363010] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 0feb5395-f648-419a-998b-eff941c16464] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.833394] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]527bc6ce-32a0-b6fd-f8a9-e9b200fa8447, 'name': SearchDatastore_Task, 'duration_secs': 0.007674} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.834167] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3f57ed6-6b05-4f66-8a2a-d7d0b3c3098f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.838839] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1380.838839] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52297f7c-aab9-767a-a04e-bd6364be9a43" [ 1380.838839] env[62235]: _type = "Task" [ 1380.838839] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.845945] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52297f7c-aab9-767a-a04e-bd6364be9a43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.866418] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ef220f90-5581-4a51-b308-12da850782a7] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.348518] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52297f7c-aab9-767a-a04e-bd6364be9a43, 'name': SearchDatastore_Task, 'duration_secs': 0.008407} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.348788] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1381.349068] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c/ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1381.349321] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89b6e529-8c78-4cb0-91b4-3328a81e9eef {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.356281] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1381.356281] env[62235]: value = "task-1272385" [ 1381.356281] env[62235]: _type = "Task" [ 1381.356281] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1381.363312] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272385, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1381.368774] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 738750ad-931e-45e4-b0c6-8232fe1b9f46] Instance has had 0 of 5 cleanup attempts {{(pid=62235) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.865620] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272385, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412729} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.865913] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c/ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1381.866122] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1381.866379] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78459d40-0ef6-4245-bd7e-509a47eb1a72 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.871954] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1381.872148] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Cleaning up deleted instances with incomplete migration {{(pid=62235) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1381.874252] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1381.874252] env[62235]: value = "task-1272386" [ 1381.874252] env[62235]: _type = "Task" [ 1381.874252] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1381.882605] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.384197] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065696} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1382.384484] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1382.385241] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f1e53e-e671-424d-b975-33a3bd736975 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.406256] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c/ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1382.406477] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea658edc-d62f-48ca-bf88-cdf70223b874 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.424874] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1382.424874] env[62235]: value = "task-1272387" [ 1382.424874] env[62235]: _type = "Task" [ 1382.424874] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1382.431860] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272387, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.935559] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272387, 'name': ReconfigVM_Task, 'duration_secs': 0.253553} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1382.935883] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Reconfigured VM instance instance-00000070 to attach disk [datastore1] ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c/ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1382.936506] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3d9ab00-1b14-452b-bd38-8aa49080c18e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.942980] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1382.942980] env[62235]: value = "task-1272388" [ 1382.942980] env[62235]: _type = "Task" [ 1382.942980] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1382.950293] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272388, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1383.453933] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272388, 'name': Rename_Task, 'duration_secs': 0.174887} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1383.454220] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1383.454438] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e804111-790d-429d-b0d8-37465dfa8a91 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.460943] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1383.460943] env[62235]: value = "task-1272389" [ 1383.460943] env[62235]: _type = "Task" [ 1383.460943] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1383.468037] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1383.971819] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272389, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1384.471635] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272389, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1384.972260] env[62235]: DEBUG oslo_vmware.api [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272389, 'name': PowerOnVM_Task, 'duration_secs': 1.058245} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1384.972623] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1384.972753] env[62235]: INFO nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Took 8.60 seconds to spawn the instance on the hypervisor. [ 1384.972963] env[62235]: DEBUG nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1384.973715] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b056853-3ca5-4504-81cf-031772273bae {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.489773] env[62235]: INFO nova.compute.manager [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Took 13.27 seconds to build instance. [ 1385.991783] env[62235]: DEBUG oslo_concurrency.lockutils [None req-6ed214f9-baf4-40da-b8a9-bb40fc22520f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.776s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1386.708361] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1386.708643] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1386.708859] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1386.709068] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1386.709256] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1386.711497] env[62235]: INFO nova.compute.manager [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Terminating instance [ 1386.713274] env[62235]: DEBUG nova.compute.manager [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1386.713472] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1386.714349] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c656ad3-4f07-4e68-b2bb-f5baaaded349 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.722230] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1386.722457] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0148d82b-999c-4331-8fbd-bdff42db0e23 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.729370] env[62235]: DEBUG oslo_vmware.api [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1386.729370] env[62235]: value = "task-1272390" [ 1386.729370] env[62235]: _type = "Task" [ 1386.729370] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1386.737225] env[62235]: DEBUG oslo_vmware.api [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272390, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.868203] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1386.868437] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1386.868562] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1386.868683] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1387.239189] env[62235]: DEBUG oslo_vmware.api [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272390, 'name': PowerOffVM_Task, 'duration_secs': 0.190775} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1387.239513] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1387.239627] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1387.239875] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-241c5a42-022e-4faa-a470-eb450107be86 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.371796] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Skipping network cache update for instance because it is being deleted. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1387.397423] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1387.397574] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1387.397724] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1387.397877] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid 8e177c90-8d68-4d70-9134-d2635abdfe7f {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1388.120706] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1388.120941] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1388.121142] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleting the datastore file [datastore1] ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1388.121425] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34cbe383-1dac-4417-9f6f-4205f829afd9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.127609] env[62235]: DEBUG oslo_vmware.api [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1388.127609] env[62235]: value = "task-1272392" [ 1388.127609] env[62235]: _type = "Task" [ 1388.127609] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.134892] env[62235]: DEBUG oslo_vmware.api [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.637459] env[62235]: DEBUG oslo_vmware.api [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141734} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1388.637794] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1388.637899] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1388.638094] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1388.638275] env[62235]: INFO nova.compute.manager [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Took 1.92 seconds to destroy the instance on the hypervisor. [ 1388.638511] env[62235]: DEBUG oslo.service.loopingcall [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1388.638706] env[62235]: DEBUG nova.compute.manager [-] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1388.638799] env[62235]: DEBUG nova.network.neutron [-] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1388.868171] env[62235]: DEBUG nova.compute.manager [req-127abfbc-00e4-44df-99b4-373cb33dce55 req-5a1c8c1b-2ce4-4fdb-848b-3fcd3c9d1fdf service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Received event network-vif-deleted-7d09f3b7-e012-44b1-81ec-5ba677f6848e {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1388.868171] env[62235]: INFO nova.compute.manager [req-127abfbc-00e4-44df-99b4-373cb33dce55 req-5a1c8c1b-2ce4-4fdb-848b-3fcd3c9d1fdf service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Neutron deleted interface 7d09f3b7-e012-44b1-81ec-5ba677f6848e; detaching it from the instance and deleting it from the info cache [ 1388.868431] env[62235]: DEBUG nova.network.neutron [req-127abfbc-00e4-44df-99b4-373cb33dce55 req-5a1c8c1b-2ce4-4fdb-848b-3fcd3c9d1fdf service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.091192] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updating instance_info_cache with network_info: [{"id": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "address": "fa:16:3e:73:ba:c1", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap909b7c34-70", "ovs_interfaceid": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.348855] env[62235]: DEBUG nova.network.neutron [-] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.371660] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e68335b3-fe43-4dec-9a6c-365e9eb3e337 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.383432] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe5803c-19d7-4e5a-8e67-6ab3e2490560 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.406529] env[62235]: DEBUG nova.compute.manager [req-127abfbc-00e4-44df-99b4-373cb33dce55 req-5a1c8c1b-2ce4-4fdb-848b-3fcd3c9d1fdf service nova] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Detach interface failed, port_id=7d09f3b7-e012-44b1-81ec-5ba677f6848e, reason: Instance ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1389.594411] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1389.594654] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1389.594952] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.595037] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.595188] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.595354] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.595519] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.595668] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.595793] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1389.595932] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.851570] env[62235]: INFO nova.compute.manager [-] [instance: ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c] Took 1.21 seconds to deallocate network for instance. [ 1390.100046] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1390.100259] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1390.100433] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1390.100600] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1390.101521] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d41555-5739-4cc6-b09b-8083ec9c8127 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.109514] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1b7bed-dfa6-4b69-88d8-f3f793c1f01d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.124171] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80be7bf-1be9-474a-aa08-51b342bb8271 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.130098] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe47cbe-f06f-45df-b2ee-307e47837328 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.158146] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181376MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1390.158285] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1390.158468] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1390.357534] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1391.214624] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 8e177c90-8d68-4d70-9134-d2635abdfe7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1391.214898] env[62235]: WARNING nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1391.215040] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1391.215145] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1391.252678] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79ce4d4-3be5-46e7-ba9a-682653b64a57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.260318] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe27327-b334-4723-ab0c-ec14127baebf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.290296] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0dc19b-82bc-4c56-b115-df8d97c67052 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.296964] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb0ebdf-63c6-471b-8a43-96a5b7e146c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.309425] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1391.813012] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1392.318692] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1392.319090] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.160s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1392.319234] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.962s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1392.319446] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1392.341398] env[62235]: INFO nova.scheduler.client.report [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted allocations for instance ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c [ 1392.849961] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5e7ce3b1-e7fa-47cc-b3dc-e3a801cf00df tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "ee7cd4ed-cf13-4ac6-88ce-b6c5266fa55c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.141s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1393.660370] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "a45e4262-c095-4616-bfcb-225755bfff5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1393.660659] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "a45e4262-c095-4616-bfcb-225755bfff5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.163447] env[62235]: DEBUG nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1394.686035] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.686312] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.687728] env[62235]: INFO nova.compute.claims [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1395.731717] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0567629-6ea1-44c0-87de-4a014d95c8eb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.739699] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407a0dee-95a3-4ef1-9332-64b9eca587c9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.768360] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7e8d66-f970-4239-be7b-3f91f7990ce4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.774968] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e20372-9a30-4253-a25e-35bebe4e2041 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.787619] env[62235]: DEBUG nova.compute.provider_tree [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1396.291159] env[62235]: DEBUG nova.scheduler.client.report [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1396.797024] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1396.797596] env[62235]: DEBUG nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1397.302649] env[62235]: DEBUG nova.compute.utils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1397.304120] env[62235]: DEBUG nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1397.304277] env[62235]: DEBUG nova.network.neutron [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1397.358198] env[62235]: DEBUG nova.policy [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8de8d8815704b1ca99fe1f48d746109', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd439d7dfff414de6ba781d2a7d464120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1397.604429] env[62235]: DEBUG nova.network.neutron [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Successfully created port: 984c876c-4643-4b32-9dad-cd09c2341ab3 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1397.808094] env[62235]: DEBUG nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1398.818748] env[62235]: DEBUG nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1398.846344] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1398.846599] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1398.846764] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1398.846957] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1398.847118] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1398.847274] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1398.847505] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1398.847709] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1398.847915] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1398.848098] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1398.848279] env[62235]: DEBUG nova.virt.hardware [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1398.849148] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f151c3c8-cf0e-4c7f-bc2a-c413fd183e46 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.856973] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ce7ec5-846b-4be8-a5fd-6ebeb2de149e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.968454] env[62235]: DEBUG nova.compute.manager [req-c6494838-58c3-40d6-8cee-328d6ce9a1ba req-fac2dc7f-0aa0-418c-9318-a52a0d1a401b service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Received event network-vif-plugged-984c876c-4643-4b32-9dad-cd09c2341ab3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1398.968612] env[62235]: DEBUG oslo_concurrency.lockutils [req-c6494838-58c3-40d6-8cee-328d6ce9a1ba req-fac2dc7f-0aa0-418c-9318-a52a0d1a401b service nova] Acquiring lock "a45e4262-c095-4616-bfcb-225755bfff5a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1398.968811] env[62235]: DEBUG oslo_concurrency.lockutils [req-c6494838-58c3-40d6-8cee-328d6ce9a1ba req-fac2dc7f-0aa0-418c-9318-a52a0d1a401b service nova] Lock "a45e4262-c095-4616-bfcb-225755bfff5a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1398.968983] env[62235]: DEBUG oslo_concurrency.lockutils [req-c6494838-58c3-40d6-8cee-328d6ce9a1ba req-fac2dc7f-0aa0-418c-9318-a52a0d1a401b service nova] Lock "a45e4262-c095-4616-bfcb-225755bfff5a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1398.969171] env[62235]: DEBUG nova.compute.manager [req-c6494838-58c3-40d6-8cee-328d6ce9a1ba req-fac2dc7f-0aa0-418c-9318-a52a0d1a401b service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] No waiting events found dispatching network-vif-plugged-984c876c-4643-4b32-9dad-cd09c2341ab3 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1398.969343] env[62235]: WARNING nova.compute.manager [req-c6494838-58c3-40d6-8cee-328d6ce9a1ba req-fac2dc7f-0aa0-418c-9318-a52a0d1a401b service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Received unexpected event network-vif-plugged-984c876c-4643-4b32-9dad-cd09c2341ab3 for instance with vm_state building and task_state spawning. [ 1399.051855] env[62235]: DEBUG nova.network.neutron [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Successfully updated port: 984c876c-4643-4b32-9dad-cd09c2341ab3 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1399.555825] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "refresh_cache-a45e4262-c095-4616-bfcb-225755bfff5a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1399.555825] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "refresh_cache-a45e4262-c095-4616-bfcb-225755bfff5a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1399.555825] env[62235]: DEBUG nova.network.neutron [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1400.086593] env[62235]: DEBUG nova.network.neutron [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1400.206170] env[62235]: DEBUG nova.network.neutron [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Updating instance_info_cache with network_info: [{"id": "984c876c-4643-4b32-9dad-cd09c2341ab3", "address": "fa:16:3e:96:a0:58", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984c876c-46", "ovs_interfaceid": "984c876c-4643-4b32-9dad-cd09c2341ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1400.708568] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "refresh_cache-a45e4262-c095-4616-bfcb-225755bfff5a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1400.708900] env[62235]: DEBUG nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Instance network_info: |[{"id": "984c876c-4643-4b32-9dad-cd09c2341ab3", "address": "fa:16:3e:96:a0:58", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984c876c-46", "ovs_interfaceid": "984c876c-4643-4b32-9dad-cd09c2341ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1400.709350] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:a0:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '849fc06e-dfc2-470f-8490-034590682ea7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '984c876c-4643-4b32-9dad-cd09c2341ab3', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1400.716655] env[62235]: DEBUG oslo.service.loopingcall [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1400.716873] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1400.717499] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58214417-a3fe-4dc0-8e9b-0c654964b38a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.736560] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1400.736560] env[62235]: value = "task-1272393" [ 1400.736560] env[62235]: _type = "Task" [ 1400.736560] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.744053] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272393, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.994359] env[62235]: DEBUG nova.compute.manager [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Received event network-changed-984c876c-4643-4b32-9dad-cd09c2341ab3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1400.994499] env[62235]: DEBUG nova.compute.manager [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Refreshing instance network info cache due to event network-changed-984c876c-4643-4b32-9dad-cd09c2341ab3. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1400.995242] env[62235]: DEBUG oslo_concurrency.lockutils [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] Acquiring lock "refresh_cache-a45e4262-c095-4616-bfcb-225755bfff5a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1400.995413] env[62235]: DEBUG oslo_concurrency.lockutils [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] Acquired lock "refresh_cache-a45e4262-c095-4616-bfcb-225755bfff5a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1400.995586] env[62235]: DEBUG nova.network.neutron [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Refreshing network info cache for port 984c876c-4643-4b32-9dad-cd09c2341ab3 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1401.245924] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272393, 'name': CreateVM_Task, 'duration_secs': 0.303494} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.246284] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1401.246730] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1401.246918] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1401.247282] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1401.247535] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc8a3d3b-6b80-4fe6-bf35-975c58ef9333 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.251799] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1401.251799] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e06e27-a1a1-ef36-2348-628ac163c7f1" [ 1401.251799] env[62235]: _type = "Task" [ 1401.251799] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.260297] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e06e27-a1a1-ef36-2348-628ac163c7f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.669238] env[62235]: DEBUG nova.network.neutron [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Updated VIF entry in instance network info cache for port 984c876c-4643-4b32-9dad-cd09c2341ab3. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1401.669606] env[62235]: DEBUG nova.network.neutron [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Updating instance_info_cache with network_info: [{"id": "984c876c-4643-4b32-9dad-cd09c2341ab3", "address": "fa:16:3e:96:a0:58", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984c876c-46", "ovs_interfaceid": "984c876c-4643-4b32-9dad-cd09c2341ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1401.762010] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52e06e27-a1a1-ef36-2348-628ac163c7f1, 'name': SearchDatastore_Task, 'duration_secs': 0.009142} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.762329] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1401.762569] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1401.762799] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1401.762952] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1401.763148] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1401.763404] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2706f995-e261-4102-9d21-34b5b7cc1ae9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.771227] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1401.771405] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1401.772077] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-deec39e5-9b35-45bd-87a7-efe0d4d5396f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.779082] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1401.779082] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dbc163-3c0f-9675-e1e6-78c812f8070b" [ 1401.779082] env[62235]: _type = "Task" [ 1401.779082] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1401.784218] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dbc163-3c0f-9675-e1e6-78c812f8070b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.172085] env[62235]: DEBUG oslo_concurrency.lockutils [req-46c326c6-35bb-4add-83b8-67986290b0e9 req-f8e7a84d-7a04-46a3-ade4-526165399fd4 service nova] Releasing lock "refresh_cache-a45e4262-c095-4616-bfcb-225755bfff5a" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1402.287619] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dbc163-3c0f-9675-e1e6-78c812f8070b, 'name': SearchDatastore_Task, 'duration_secs': 0.00797} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1402.288402] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-154db0f5-7b08-4314-9bb4-425ede2f3934 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.293331] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1402.293331] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5261edf2-5184-1a12-1a61-a175a31e8df5" [ 1402.293331] env[62235]: _type = "Task" [ 1402.293331] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.300488] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5261edf2-5184-1a12-1a61-a175a31e8df5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1402.803908] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5261edf2-5184-1a12-1a61-a175a31e8df5, 'name': SearchDatastore_Task, 'duration_secs': 0.009661} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1402.804193] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1402.804457] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a45e4262-c095-4616-bfcb-225755bfff5a/a45e4262-c095-4616-bfcb-225755bfff5a.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1402.804739] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64a5bdd8-cf0c-4b79-87b0-e6e64edcb866 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.811031] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1402.811031] env[62235]: value = "task-1272394" [ 1402.811031] env[62235]: _type = "Task" [ 1402.811031] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.818654] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272394, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.320927] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272394, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496884} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1403.321366] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] a45e4262-c095-4616-bfcb-225755bfff5a/a45e4262-c095-4616-bfcb-225755bfff5a.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1403.321441] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1403.321648] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71906ebf-6ef2-490d-b592-aaf8f217d8a6 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.328273] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1403.328273] env[62235]: value = "task-1272395" [ 1403.328273] env[62235]: _type = "Task" [ 1403.328273] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.335886] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272395, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.837328] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272395, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061126} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1403.837598] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1403.838348] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb72432-9a22-45ea-a80c-369536699fb7 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.859036] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] a45e4262-c095-4616-bfcb-225755bfff5a/a45e4262-c095-4616-bfcb-225755bfff5a.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1403.859269] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01ef339c-1b26-4de4-a8da-50a3f78b7faf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.877101] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1403.877101] env[62235]: value = "task-1272396" [ 1403.877101] env[62235]: _type = "Task" [ 1403.877101] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.885402] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.387325] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272396, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.888090] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272396, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.389079] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272396, 'name': ReconfigVM_Task, 'duration_secs': 1.218323} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.389464] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Reconfigured VM instance instance-00000071 to attach disk [datastore2] a45e4262-c095-4616-bfcb-225755bfff5a/a45e4262-c095-4616-bfcb-225755bfff5a.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1405.389972] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5721610f-1648-48b9-8b0c-d3db6fa6e09d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.396379] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1405.396379] env[62235]: value = "task-1272397" [ 1405.396379] env[62235]: _type = "Task" [ 1405.396379] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.403704] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272397, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.906225] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272397, 'name': Rename_Task, 'duration_secs': 0.132195} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.906508] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1405.906754] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca4bddec-b379-4e9e-9f3f-56bbfd982151 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.913300] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1405.913300] env[62235]: value = "task-1272398" [ 1405.913300] env[62235]: _type = "Task" [ 1405.913300] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.920622] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272398, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.423521] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272398, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.923793] env[62235]: DEBUG oslo_vmware.api [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272398, 'name': PowerOnVM_Task, 'duration_secs': 0.692633} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.924085] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1406.924297] env[62235]: INFO nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Took 8.11 seconds to spawn the instance on the hypervisor. [ 1406.924479] env[62235]: DEBUG nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1406.925249] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d46be4f-ce57-4d76-addd-31786aa210e0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.443148] env[62235]: INFO nova.compute.manager [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Took 12.77 seconds to build instance. [ 1407.945316] env[62235]: DEBUG oslo_concurrency.lockutils [None req-5307625f-bf47-4559-bc7c-a2869a7810f4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "a45e4262-c095-4616-bfcb-225755bfff5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.284s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1408.214286] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8b32a6dd-2733-4e9a-a436-9f24988189ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1408.214525] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1408.717624] env[62235]: DEBUG nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1409.240420] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1409.240797] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1409.243078] env[62235]: INFO nova.compute.claims [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1410.296335] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6834a8-a5b6-48dd-b2c0-ab8a555c1160 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.304013] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07874bbc-8273-4a59-b84a-1636290595cf {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.332810] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919a517f-6e3e-4b39-8834-a6edd9b5bb69 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.339502] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23dd1b23-7359-4eba-939a-bd744e33ef8a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.352177] env[62235]: DEBUG nova.compute.provider_tree [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1410.854833] env[62235]: DEBUG nova.scheduler.client.report [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1411.360695] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1411.361282] env[62235]: DEBUG nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1411.866228] env[62235]: DEBUG nova.compute.utils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1411.867692] env[62235]: DEBUG nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1411.867874] env[62235]: DEBUG nova.network.neutron [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1411.919787] env[62235]: DEBUG nova.policy [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8de8d8815704b1ca99fe1f48d746109', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd439d7dfff414de6ba781d2a7d464120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1412.161246] env[62235]: DEBUG nova.network.neutron [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Successfully created port: cfa27519-b47b-4da7-9053-ce5352b0630b {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1412.371831] env[62235]: DEBUG nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1413.384563] env[62235]: DEBUG nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1413.410874] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1413.411059] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1413.411128] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1413.411270] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1413.411450] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1413.411570] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1413.411780] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1413.412293] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1413.412293] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1413.412293] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1413.412460] env[62235]: DEBUG nova.virt.hardware [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1413.413306] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b49739a-8bc6-458c-8700-8e7e32b89743 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.421295] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fabf82-f153-4328-baf2-ab28cb8137f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.546755] env[62235]: DEBUG nova.compute.manager [req-fd5afe03-574b-4484-a49d-48d5294b3625 req-22d039cc-710f-4066-bd6f-ec26470f4e04 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Received event network-vif-plugged-cfa27519-b47b-4da7-9053-ce5352b0630b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1413.547075] env[62235]: DEBUG oslo_concurrency.lockutils [req-fd5afe03-574b-4484-a49d-48d5294b3625 req-22d039cc-710f-4066-bd6f-ec26470f4e04 service nova] Acquiring lock "8b32a6dd-2733-4e9a-a436-9f24988189ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1413.547239] env[62235]: DEBUG oslo_concurrency.lockutils [req-fd5afe03-574b-4484-a49d-48d5294b3625 req-22d039cc-710f-4066-bd6f-ec26470f4e04 service nova] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1413.547392] env[62235]: DEBUG oslo_concurrency.lockutils [req-fd5afe03-574b-4484-a49d-48d5294b3625 req-22d039cc-710f-4066-bd6f-ec26470f4e04 service nova] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1413.547660] env[62235]: DEBUG nova.compute.manager [req-fd5afe03-574b-4484-a49d-48d5294b3625 req-22d039cc-710f-4066-bd6f-ec26470f4e04 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] No waiting events found dispatching network-vif-plugged-cfa27519-b47b-4da7-9053-ce5352b0630b {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1413.547864] env[62235]: WARNING nova.compute.manager [req-fd5afe03-574b-4484-a49d-48d5294b3625 req-22d039cc-710f-4066-bd6f-ec26470f4e04 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Received unexpected event network-vif-plugged-cfa27519-b47b-4da7-9053-ce5352b0630b for instance with vm_state building and task_state spawning. [ 1414.076382] env[62235]: DEBUG nova.network.neutron [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Successfully updated port: cfa27519-b47b-4da7-9053-ce5352b0630b {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1414.099651] env[62235]: DEBUG nova.compute.manager [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Received event network-changed-cfa27519-b47b-4da7-9053-ce5352b0630b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1414.099783] env[62235]: DEBUG nova.compute.manager [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Refreshing instance network info cache due to event network-changed-cfa27519-b47b-4da7-9053-ce5352b0630b. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1414.100041] env[62235]: DEBUG oslo_concurrency.lockutils [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] Acquiring lock "refresh_cache-8b32a6dd-2733-4e9a-a436-9f24988189ea" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1414.100146] env[62235]: DEBUG oslo_concurrency.lockutils [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] Acquired lock "refresh_cache-8b32a6dd-2733-4e9a-a436-9f24988189ea" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1414.100313] env[62235]: DEBUG nova.network.neutron [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Refreshing network info cache for port cfa27519-b47b-4da7-9053-ce5352b0630b {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1414.578625] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "refresh_cache-8b32a6dd-2733-4e9a-a436-9f24988189ea" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1414.630599] env[62235]: DEBUG nova.network.neutron [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1414.696995] env[62235]: DEBUG nova.network.neutron [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1415.199983] env[62235]: DEBUG oslo_concurrency.lockutils [req-65625c90-296a-4963-8a4c-1dd9d269b150 req-9ea1ac54-64a8-419a-8ff6-5b1bb4929a34 service nova] Releasing lock "refresh_cache-8b32a6dd-2733-4e9a-a436-9f24988189ea" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1415.200535] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "refresh_cache-8b32a6dd-2733-4e9a-a436-9f24988189ea" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1415.200727] env[62235]: DEBUG nova.network.neutron [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1415.732187] env[62235]: DEBUG nova.network.neutron [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1415.858459] env[62235]: DEBUG nova.network.neutron [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Updating instance_info_cache with network_info: [{"id": "cfa27519-b47b-4da7-9053-ce5352b0630b", "address": "fa:16:3e:c5:12:0b", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfa27519-b4", "ovs_interfaceid": "cfa27519-b47b-4da7-9053-ce5352b0630b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1416.360767] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "refresh_cache-8b32a6dd-2733-4e9a-a436-9f24988189ea" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1416.361122] env[62235]: DEBUG nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Instance network_info: |[{"id": "cfa27519-b47b-4da7-9053-ce5352b0630b", "address": "fa:16:3e:c5:12:0b", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfa27519-b4", "ovs_interfaceid": "cfa27519-b47b-4da7-9053-ce5352b0630b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1416.361579] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:12:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '849fc06e-dfc2-470f-8490-034590682ea7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfa27519-b47b-4da7-9053-ce5352b0630b', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1416.368942] env[62235]: DEBUG oslo.service.loopingcall [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1416.369171] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1416.369394] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89ad0bcd-6d64-482d-b9f2-3b5970166980 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.389868] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1416.389868] env[62235]: value = "task-1272399" [ 1416.389868] env[62235]: _type = "Task" [ 1416.389868] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.398357] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272399, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.900064] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272399, 'name': CreateVM_Task, 'duration_secs': 0.294751} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.900422] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1416.900875] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1416.901058] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1416.901384] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1416.901631] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d71c4ca0-2dda-4b38-9d34-447825c945a5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.905984] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1416.905984] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ab7491-02df-a8dc-e537-88c37131f086" [ 1416.905984] env[62235]: _type = "Task" [ 1416.905984] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.913164] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ab7491-02df-a8dc-e537-88c37131f086, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.416465] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52ab7491-02df-a8dc-e537-88c37131f086, 'name': SearchDatastore_Task, 'duration_secs': 0.009658} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.416779] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1417.417034] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1417.417366] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1417.417530] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1417.417718] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1417.417980] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ceaa591-3203-4bc5-9f3b-81db25380fda {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.426016] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1417.426220] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1417.426898] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88af5033-cd79-406f-825d-42d2fa95cdb1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.431549] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1417.431549] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52481406-4707-9b9b-dcc8-f2c8ee86a77b" [ 1417.431549] env[62235]: _type = "Task" [ 1417.431549] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.438813] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52481406-4707-9b9b-dcc8-f2c8ee86a77b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.941472] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52481406-4707-9b9b-dcc8-f2c8ee86a77b, 'name': SearchDatastore_Task, 'duration_secs': 0.007486} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.942259] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0dc7f74-3140-4eb9-9bfb-9d8ee3ed204d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.947893] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1417.947893] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521e3e35-412e-a93c-d4df-e02dcc96dac5" [ 1417.947893] env[62235]: _type = "Task" [ 1417.947893] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.955391] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521e3e35-412e-a93c-d4df-e02dcc96dac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.458203] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521e3e35-412e-a93c-d4df-e02dcc96dac5, 'name': SearchDatastore_Task, 'duration_secs': 0.00878} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.458432] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1418.458690] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8b32a6dd-2733-4e9a-a436-9f24988189ea/8b32a6dd-2733-4e9a-a436-9f24988189ea.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1418.458939] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35365bc1-a1a7-4097-8cf7-8e3cfba62bf4 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.464898] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1418.464898] env[62235]: value = "task-1272400" [ 1418.464898] env[62235]: _type = "Task" [ 1418.464898] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.471947] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272400, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.974396] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272400, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.392353} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.974755] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] 8b32a6dd-2733-4e9a-a436-9f24988189ea/8b32a6dd-2733-4e9a-a436-9f24988189ea.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1418.974909] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1418.975175] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4452b998-9306-48c6-b6f6-5ce783f4546c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.981827] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1418.981827] env[62235]: value = "task-1272401" [ 1418.981827] env[62235]: _type = "Task" [ 1418.981827] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.988665] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272401, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1419.491483] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272401, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065894} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1419.491822] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1419.492544] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1f124a-f7d3-4cb8-b725-fd8ceeca20f1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.513710] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 8b32a6dd-2733-4e9a-a436-9f24988189ea/8b32a6dd-2733-4e9a-a436-9f24988189ea.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1419.513994] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a76cd6b3-5dbf-47f6-b0d4-b58b789f2415 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.533346] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1419.533346] env[62235]: value = "task-1272402" [ 1419.533346] env[62235]: _type = "Task" [ 1419.533346] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1419.540944] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272402, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.044789] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272402, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.544774] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272402, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.046020] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272402, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.549518] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272402, 'name': ReconfigVM_Task, 'duration_secs': 1.551264} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.549810] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 8b32a6dd-2733-4e9a-a436-9f24988189ea/8b32a6dd-2733-4e9a-a436-9f24988189ea.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1421.550476] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a095752-f72f-4158-aa17-8ecfe0cc072b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.557651] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1421.557651] env[62235]: value = "task-1272403" [ 1421.557651] env[62235]: _type = "Task" [ 1421.557651] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.566587] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272403, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1422.068754] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272403, 'name': Rename_Task, 'duration_secs': 0.148911} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1422.069086] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1422.069327] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b0d35c5-5dd1-48d9-abf3-efb35ec96471 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.076088] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1422.076088] env[62235]: value = "task-1272404" [ 1422.076088] env[62235]: _type = "Task" [ 1422.076088] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1422.087809] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1422.585640] env[62235]: DEBUG oslo_vmware.api [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272404, 'name': PowerOnVM_Task, 'duration_secs': 0.416187} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1422.586050] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1422.586306] env[62235]: INFO nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Took 9.20 seconds to spawn the instance on the hypervisor. [ 1422.586498] env[62235]: DEBUG nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1422.587273] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6651c96-8b2a-4c08-95e4-578e064cd219 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.107631] env[62235]: INFO nova.compute.manager [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Took 13.88 seconds to build instance. [ 1423.609797] env[62235]: DEBUG oslo_concurrency.lockutils [None req-16ab230f-9ee9-43ee-a9ad-84fc43fe32a9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.395s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.853478] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8b32a6dd-2733-4e9a-a436-9f24988189ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.853770] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.853985] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8b32a6dd-2733-4e9a-a436-9f24988189ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.854193] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.854362] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.856838] env[62235]: INFO nova.compute.manager [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Terminating instance [ 1423.858559] env[62235]: DEBUG nova.compute.manager [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1423.858763] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1423.859593] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5548e4-079e-4038-b8de-f9854638c5f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.866969] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1423.867202] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-90125d89-2b75-46ca-9bc1-4624a6ee3817 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.873365] env[62235]: DEBUG oslo_vmware.api [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1423.873365] env[62235]: value = "task-1272405" [ 1423.873365] env[62235]: _type = "Task" [ 1423.873365] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.881013] env[62235]: DEBUG oslo_vmware.api [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.383211] env[62235]: DEBUG oslo_vmware.api [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272405, 'name': PowerOffVM_Task, 'duration_secs': 0.167566} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.383584] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1424.383704] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1424.383895] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ca24cad-27aa-404c-b54c-bb577dc4ceb2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.480405] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1424.480624] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1424.480816] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleting the datastore file [datastore1] 8b32a6dd-2733-4e9a-a436-9f24988189ea {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1424.481091] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96639501-6bfa-4079-99d9-403204c48304 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.488372] env[62235]: DEBUG oslo_vmware.api [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1424.488372] env[62235]: value = "task-1272407" [ 1424.488372] env[62235]: _type = "Task" [ 1424.488372] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.497848] env[62235]: DEBUG oslo_vmware.api [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.998883] env[62235]: DEBUG oslo_vmware.api [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118908} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.999141] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1424.999336] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1424.999517] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1424.999695] env[62235]: INFO nova.compute.manager [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1424.999939] env[62235]: DEBUG oslo.service.loopingcall [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1425.000149] env[62235]: DEBUG nova.compute.manager [-] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1425.000247] env[62235]: DEBUG nova.network.neutron [-] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1425.288620] env[62235]: DEBUG nova.compute.manager [req-3f2aa799-289d-45a0-9853-ad378e44acde req-0b1dfffc-d00f-4212-b540-37cb2323dfe7 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Received event network-vif-deleted-cfa27519-b47b-4da7-9053-ce5352b0630b {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1425.288830] env[62235]: INFO nova.compute.manager [req-3f2aa799-289d-45a0-9853-ad378e44acde req-0b1dfffc-d00f-4212-b540-37cb2323dfe7 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Neutron deleted interface cfa27519-b47b-4da7-9053-ce5352b0630b; detaching it from the instance and deleting it from the info cache [ 1425.289013] env[62235]: DEBUG nova.network.neutron [req-3f2aa799-289d-45a0-9853-ad378e44acde req-0b1dfffc-d00f-4212-b540-37cb2323dfe7 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1425.761210] env[62235]: DEBUG nova.network.neutron [-] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1425.791385] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc51da10-70e1-43d7-bdd0-e2fe90311efe {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.801556] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6b5923-c27b-44ac-930f-677389c1e34b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.824662] env[62235]: DEBUG nova.compute.manager [req-3f2aa799-289d-45a0-9853-ad378e44acde req-0b1dfffc-d00f-4212-b540-37cb2323dfe7 service nova] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Detach interface failed, port_id=cfa27519-b47b-4da7-9053-ce5352b0630b, reason: Instance 8b32a6dd-2733-4e9a-a436-9f24988189ea could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1426.264319] env[62235]: INFO nova.compute.manager [-] [instance: 8b32a6dd-2733-4e9a-a436-9f24988189ea] Took 1.26 seconds to deallocate network for instance. [ 1426.770647] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1426.770975] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1426.771177] env[62235]: DEBUG nova.objects.instance [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lazy-loading 'resources' on Instance uuid 8b32a6dd-2733-4e9a-a436-9f24988189ea {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1427.327149] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17806f67-cb8a-482d-bc7a-b8749a7f4df5 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.334601] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec29149-1747-4304-84e2-5549c20933e1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.365148] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41572a09-c9c8-424b-89bc-364a512c2c61 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.372017] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7348f3d-ae91-4359-ac53-f9aebd98cb97 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.384910] env[62235]: DEBUG nova.compute.provider_tree [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1427.887869] env[62235]: DEBUG nova.scheduler.client.report [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1428.392596] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1428.410027] env[62235]: INFO nova.scheduler.client.report [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted allocations for instance 8b32a6dd-2733-4e9a-a436-9f24988189ea [ 1428.917791] env[62235]: DEBUG oslo_concurrency.lockutils [None req-9add8678-306e-487f-a5d9-4dec155711cd tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8b32a6dd-2733-4e9a-a436-9f24988189ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.063s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1429.334810] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "a45e4262-c095-4616-bfcb-225755bfff5a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1429.335116] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "a45e4262-c095-4616-bfcb-225755bfff5a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1429.335384] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "a45e4262-c095-4616-bfcb-225755bfff5a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1429.335586] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "a45e4262-c095-4616-bfcb-225755bfff5a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1429.335756] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "a45e4262-c095-4616-bfcb-225755bfff5a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1429.337929] env[62235]: INFO nova.compute.manager [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Terminating instance [ 1429.339694] env[62235]: DEBUG nova.compute.manager [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1429.339905] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1429.340766] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0798f422-8543-4795-9905-46fa703c3e87 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.348546] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1429.349037] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94e61ff0-ebe6-4c0a-a025-091d1a50375e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.355349] env[62235]: DEBUG oslo_vmware.api [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1429.355349] env[62235]: value = "task-1272408" [ 1429.355349] env[62235]: _type = "Task" [ 1429.355349] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.362688] env[62235]: DEBUG oslo_vmware.api [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272408, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.865856] env[62235]: DEBUG oslo_vmware.api [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272408, 'name': PowerOffVM_Task, 'duration_secs': 0.151001} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.865856] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1429.865856] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1429.866206] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c08c4621-9225-4652-b9c4-dfbc1c411861 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.026485] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1430.026837] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1430.026918] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleting the datastore file [datastore2] a45e4262-c095-4616-bfcb-225755bfff5a {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1430.027197] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5019773-6720-4889-9ac3-110f4462383d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.033642] env[62235]: DEBUG oslo_vmware.api [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1430.033642] env[62235]: value = "task-1272410" [ 1430.033642] env[62235]: _type = "Task" [ 1430.033642] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.041288] env[62235]: DEBUG oslo_vmware.api [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.544248] env[62235]: DEBUG oslo_vmware.api [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14813} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.544462] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1430.544652] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1430.544839] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1430.545025] env[62235]: INFO nova.compute.manager [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1430.545315] env[62235]: DEBUG oslo.service.loopingcall [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1430.545549] env[62235]: DEBUG nova.compute.manager [-] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1430.545650] env[62235]: DEBUG nova.network.neutron [-] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1430.817456] env[62235]: DEBUG nova.compute.manager [req-cdb7f25c-58bc-45c1-b12e-721dac361df2 req-e0117c5d-003f-4c83-92ca-b34f0d9034d0 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Received event network-vif-deleted-984c876c-4643-4b32-9dad-cd09c2341ab3 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1430.817744] env[62235]: INFO nova.compute.manager [req-cdb7f25c-58bc-45c1-b12e-721dac361df2 req-e0117c5d-003f-4c83-92ca-b34f0d9034d0 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Neutron deleted interface 984c876c-4643-4b32-9dad-cd09c2341ab3; detaching it from the instance and deleting it from the info cache [ 1430.817971] env[62235]: DEBUG nova.network.neutron [req-cdb7f25c-58bc-45c1-b12e-721dac361df2 req-e0117c5d-003f-4c83-92ca-b34f0d9034d0 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1431.272574] env[62235]: DEBUG nova.network.neutron [-] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1431.320644] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed445961-26b7-49c0-9c0d-c24fef2cd165 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.330130] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37bd250c-87cb-4366-9ccb-827f29edb07a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.351277] env[62235]: DEBUG nova.compute.manager [req-cdb7f25c-58bc-45c1-b12e-721dac361df2 req-e0117c5d-003f-4c83-92ca-b34f0d9034d0 service nova] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Detach interface failed, port_id=984c876c-4643-4b32-9dad-cd09c2341ab3, reason: Instance a45e4262-c095-4616-bfcb-225755bfff5a could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1431.775421] env[62235]: INFO nova.compute.manager [-] [instance: a45e4262-c095-4616-bfcb-225755bfff5a] Took 1.23 seconds to deallocate network for instance. [ 1432.281664] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1432.281950] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1432.282211] env[62235]: DEBUG nova.objects.instance [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lazy-loading 'resources' on Instance uuid a45e4262-c095-4616-bfcb-225755bfff5a {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1432.825624] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11568ba-958e-4bcb-9507-8cf0b2fa1321 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.833124] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12615fc-b7c3-4c42-99e8-4e370efbb386 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.863071] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a909676c-1ec3-4cf5-9073-ba30dc909913 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.869910] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328ee602-3ef3-4cb0-bc7f-ad71eb119d37 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.882684] env[62235]: DEBUG nova.compute.provider_tree [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1433.386205] env[62235]: DEBUG nova.scheduler.client.report [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1433.891565] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1433.909392] env[62235]: INFO nova.scheduler.client.report [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted allocations for instance a45e4262-c095-4616-bfcb-225755bfff5a [ 1434.416507] env[62235]: DEBUG oslo_concurrency.lockutils [None req-0d12611c-6101-40c0-be11-9a7f3aa8565f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "a45e4262-c095-4616-bfcb-225755bfff5a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.081s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1435.178130] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1435.178392] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1435.680925] env[62235]: DEBUG nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1436.199654] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1436.199915] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1436.201413] env[62235]: INFO nova.compute.claims [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1437.244564] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ec11b6-d5ff-4e15-a1fe-35a4b7871912 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.253345] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca91e42-c22d-4f73-85f8-3fea5dc97630 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.282176] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff71f0c-b41f-47bd-b686-3967283e092c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.288937] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2525c617-fc81-4bae-ade0-ec6136bd58f3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.301543] env[62235]: DEBUG nova.compute.provider_tree [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1437.805115] env[62235]: DEBUG nova.scheduler.client.report [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1438.310523] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1438.311068] env[62235]: DEBUG nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1438.816633] env[62235]: DEBUG nova.compute.utils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1438.818177] env[62235]: DEBUG nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1438.818286] env[62235]: DEBUG nova.network.neutron [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1438.876774] env[62235]: DEBUG nova.policy [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8de8d8815704b1ca99fe1f48d746109', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd439d7dfff414de6ba781d2a7d464120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1439.130668] env[62235]: DEBUG nova.network.neutron [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Successfully created port: 3bfe8ae8-f413-45d7-9899-1f459beb7534 {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1439.321200] env[62235]: DEBUG nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1440.330902] env[62235]: DEBUG nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1440.356012] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1440.356350] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1440.356555] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1440.356761] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1440.356916] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1440.357080] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1440.357296] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1440.357463] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1440.357640] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1440.357828] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1440.358037] env[62235]: DEBUG nova.virt.hardware [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1440.358917] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6366974-60d0-439e-87d8-e9050621e9ee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.367271] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbedcfbb-f00a-44c1-96f5-73f019fb309b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.508200] env[62235]: DEBUG nova.compute.manager [req-9016d3a9-7e49-45da-a662-5a70c8178175 req-3e4b7081-8741-41e8-ace0-4a512895e93d service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Received event network-vif-plugged-3bfe8ae8-f413-45d7-9899-1f459beb7534 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1440.508435] env[62235]: DEBUG oslo_concurrency.lockutils [req-9016d3a9-7e49-45da-a662-5a70c8178175 req-3e4b7081-8741-41e8-ace0-4a512895e93d service nova] Acquiring lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1440.508654] env[62235]: DEBUG oslo_concurrency.lockutils [req-9016d3a9-7e49-45da-a662-5a70c8178175 req-3e4b7081-8741-41e8-ace0-4a512895e93d service nova] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1440.508826] env[62235]: DEBUG oslo_concurrency.lockutils [req-9016d3a9-7e49-45da-a662-5a70c8178175 req-3e4b7081-8741-41e8-ace0-4a512895e93d service nova] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1440.508998] env[62235]: DEBUG nova.compute.manager [req-9016d3a9-7e49-45da-a662-5a70c8178175 req-3e4b7081-8741-41e8-ace0-4a512895e93d service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] No waiting events found dispatching network-vif-plugged-3bfe8ae8-f413-45d7-9899-1f459beb7534 {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1440.509185] env[62235]: WARNING nova.compute.manager [req-9016d3a9-7e49-45da-a662-5a70c8178175 req-3e4b7081-8741-41e8-ace0-4a512895e93d service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Received unexpected event network-vif-plugged-3bfe8ae8-f413-45d7-9899-1f459beb7534 for instance with vm_state building and task_state spawning. [ 1440.590138] env[62235]: DEBUG nova.network.neutron [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Successfully updated port: 3bfe8ae8-f413-45d7-9899-1f459beb7534 {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1441.092653] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "refresh_cache-e3fc3ba2-5081-455f-be58-fd8b784c4db8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1441.092823] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "refresh_cache-e3fc3ba2-5081-455f-be58-fd8b784c4db8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1441.092985] env[62235]: DEBUG nova.network.neutron [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1441.628279] env[62235]: DEBUG nova.network.neutron [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1441.748027] env[62235]: DEBUG nova.network.neutron [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Updating instance_info_cache with network_info: [{"id": "3bfe8ae8-f413-45d7-9899-1f459beb7534", "address": "fa:16:3e:8e:fb:f0", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bfe8ae8-f4", "ovs_interfaceid": "3bfe8ae8-f413-45d7-9899-1f459beb7534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1442.250713] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "refresh_cache-e3fc3ba2-5081-455f-be58-fd8b784c4db8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1442.251065] env[62235]: DEBUG nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Instance network_info: |[{"id": "3bfe8ae8-f413-45d7-9899-1f459beb7534", "address": "fa:16:3e:8e:fb:f0", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bfe8ae8-f4", "ovs_interfaceid": "3bfe8ae8-f413-45d7-9899-1f459beb7534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1442.251506] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:fb:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '849fc06e-dfc2-470f-8490-034590682ea7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bfe8ae8-f413-45d7-9899-1f459beb7534', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1442.258839] env[62235]: DEBUG oslo.service.loopingcall [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1442.259066] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1442.259300] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76cd7933-edcb-438d-ba31-38d48e740048 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.279296] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1442.279296] env[62235]: value = "task-1272411" [ 1442.279296] env[62235]: _type = "Task" [ 1442.279296] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.287240] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272411, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1442.535949] env[62235]: DEBUG nova.compute.manager [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Received event network-changed-3bfe8ae8-f413-45d7-9899-1f459beb7534 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1442.536176] env[62235]: DEBUG nova.compute.manager [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Refreshing instance network info cache due to event network-changed-3bfe8ae8-f413-45d7-9899-1f459beb7534. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1442.536398] env[62235]: DEBUG oslo_concurrency.lockutils [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] Acquiring lock "refresh_cache-e3fc3ba2-5081-455f-be58-fd8b784c4db8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1442.536592] env[62235]: DEBUG oslo_concurrency.lockutils [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] Acquired lock "refresh_cache-e3fc3ba2-5081-455f-be58-fd8b784c4db8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1442.536773] env[62235]: DEBUG nova.network.neutron [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Refreshing network info cache for port 3bfe8ae8-f413-45d7-9899-1f459beb7534 {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1442.789796] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272411, 'name': CreateVM_Task, 'duration_secs': 0.262903} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.790153] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1442.790602] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1442.790775] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1442.791103] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1442.791346] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdfa8ead-01ce-4d04-ac19-d9727c17f08e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.795555] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1442.795555] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a294aa-dafb-6e23-bd29-b1ea8e2d107d" [ 1442.795555] env[62235]: _type = "Task" [ 1442.795555] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.802928] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a294aa-dafb-6e23-bd29-b1ea8e2d107d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1443.209348] env[62235]: DEBUG nova.network.neutron [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Updated VIF entry in instance network info cache for port 3bfe8ae8-f413-45d7-9899-1f459beb7534. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1443.209704] env[62235]: DEBUG nova.network.neutron [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Updating instance_info_cache with network_info: [{"id": "3bfe8ae8-f413-45d7-9899-1f459beb7534", "address": "fa:16:3e:8e:fb:f0", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bfe8ae8-f4", "ovs_interfaceid": "3bfe8ae8-f413-45d7-9899-1f459beb7534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1443.305879] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52a294aa-dafb-6e23-bd29-b1ea8e2d107d, 'name': SearchDatastore_Task, 'duration_secs': 0.008646} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1443.306209] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1443.306445] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1443.306688] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1443.307198] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1443.307198] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1443.307311] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30675a21-cc74-4a1b-a021-b12eec7f6268 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.314999] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1443.315160] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1443.315849] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39e2947d-c2fd-486b-b4a9-e5aa7e16bff0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.320243] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1443.320243] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521ad13a-5056-13a2-1706-29bd71a8f576" [ 1443.320243] env[62235]: _type = "Task" [ 1443.320243] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1443.327209] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521ad13a-5056-13a2-1706-29bd71a8f576, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1443.712271] env[62235]: DEBUG oslo_concurrency.lockutils [req-a65e5937-0db4-4261-a1a5-f4d6871e9fc1 req-8e1097e9-8ca3-4b62-add6-1c547029b544 service nova] Releasing lock "refresh_cache-e3fc3ba2-5081-455f-be58-fd8b784c4db8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1443.830906] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]521ad13a-5056-13a2-1706-29bd71a8f576, 'name': SearchDatastore_Task, 'duration_secs': 0.007623} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1443.831696] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a2b9370-5de2-452a-ac89-5bb42b894d3e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.836479] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1443.836479] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524c5324-cd7e-0ab6-ac0c-42feb02de34d" [ 1443.836479] env[62235]: _type = "Task" [ 1443.836479] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1443.843923] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524c5324-cd7e-0ab6-ac0c-42feb02de34d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1444.346519] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]524c5324-cd7e-0ab6-ac0c-42feb02de34d, 'name': SearchDatastore_Task, 'duration_secs': 0.008403} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1444.346731] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1444.346984] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] e3fc3ba2-5081-455f-be58-fd8b784c4db8/e3fc3ba2-5081-455f-be58-fd8b784c4db8.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1444.347253] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d32485d-d1c3-4f43-9977-5787fb5e800e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.354111] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1444.354111] env[62235]: value = "task-1272412" [ 1444.354111] env[62235]: _type = "Task" [ 1444.354111] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1444.361740] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1444.864304] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.398738} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1444.864654] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore1] e3fc3ba2-5081-455f-be58-fd8b784c4db8/e3fc3ba2-5081-455f-be58-fd8b784c4db8.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1444.864727] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1444.864981] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-35a7d774-5434-446f-8d79-4ae10bf32282 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.871159] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1444.871159] env[62235]: value = "task-1272413" [ 1444.871159] env[62235]: _type = "Task" [ 1444.871159] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1444.877996] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272413, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1445.381275] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272413, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061881} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1445.381545] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1445.382314] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cee410e-fe2d-4362-bcd1-6398c4201eee {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.403314] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] e3fc3ba2-5081-455f-be58-fd8b784c4db8/e3fc3ba2-5081-455f-be58-fd8b784c4db8.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1445.403546] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a873325-3781-4031-b358-64a89aa4976b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.423818] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1445.423818] env[62235]: value = "task-1272414" [ 1445.423818] env[62235]: _type = "Task" [ 1445.423818] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.431183] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272414, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1445.933785] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272414, 'name': ReconfigVM_Task, 'duration_secs': 0.261293} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1445.934203] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Reconfigured VM instance instance-00000073 to attach disk [datastore1] e3fc3ba2-5081-455f-be58-fd8b784c4db8/e3fc3ba2-5081-455f-be58-fd8b784c4db8.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1445.934712] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a3d8647-8d5c-4e16-920e-782d56fd9063 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.940293] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1445.940293] env[62235]: value = "task-1272415" [ 1445.940293] env[62235]: _type = "Task" [ 1445.940293] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.948584] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272415, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1446.449901] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272415, 'name': Rename_Task, 'duration_secs': 0.133859} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1446.450192] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1446.450437] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-163da747-5b14-45b9-9ddc-8b2fd636c144 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.455948] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1446.455948] env[62235]: value = "task-1272416" [ 1446.455948] env[62235]: _type = "Task" [ 1446.455948] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1446.463098] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1446.966580] env[62235]: DEBUG oslo_vmware.api [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272416, 'name': PowerOnVM_Task, 'duration_secs': 0.421795} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1446.969034] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1446.969034] env[62235]: INFO nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1446.969034] env[62235]: DEBUG nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1446.969034] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2031ea-f236-42bf-8a74-33caf8837298 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1447.485828] env[62235]: INFO nova.compute.manager [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Took 11.30 seconds to build instance. [ 1447.768898] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1447.769169] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1447.987882] env[62235]: DEBUG oslo_concurrency.lockutils [None req-b323198c-e403-4f65-bb16-86f4bd51f41f tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.809s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1448.273515] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1448.273647] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Starting heal instance info cache {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1448.273749] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Rebuilding the list of instances to heal {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1448.587901] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.588246] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1448.588464] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1448.588650] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1448.588824] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1448.591190] env[62235]: INFO nova.compute.manager [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Terminating instance [ 1448.592929] env[62235]: DEBUG nova.compute.manager [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1448.593144] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1448.594321] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78e5ad7-7130-4e27-9680-8dd292f2a124 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.601614] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1448.601833] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f21814db-8cce-4fc6-ac01-76554d7a1e81 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.607859] env[62235]: DEBUG oslo_vmware.api [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1448.607859] env[62235]: value = "task-1272417" [ 1448.607859] env[62235]: _type = "Task" [ 1448.607859] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1448.615225] env[62235]: DEBUG oslo_vmware.api [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1448.777802] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Skipping network cache update for instance because it is being deleted. {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1448.804667] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1448.804819] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquired lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1448.804966] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Forcefully refreshing network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1448.805138] env[62235]: DEBUG nova.objects.instance [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lazy-loading 'info_cache' on Instance uuid 8e177c90-8d68-4d70-9134-d2635abdfe7f {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1449.117418] env[62235]: DEBUG oslo_vmware.api [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272417, 'name': PowerOffVM_Task, 'duration_secs': 0.172915} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1449.117786] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1449.117935] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1449.118138] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc6f9afa-ebbd-4961-87a7-b6e9f45fb106 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.213075] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1449.213263] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Deleting contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1449.213439] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleting the datastore file [datastore1] e3fc3ba2-5081-455f-be58-fd8b784c4db8 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1449.213746] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8315087-c8d8-4153-bb22-51b101f3e1f2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1449.219969] env[62235]: DEBUG oslo_vmware.api [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1449.219969] env[62235]: value = "task-1272419" [ 1449.219969] env[62235]: _type = "Task" [ 1449.219969] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1449.227524] env[62235]: DEBUG oslo_vmware.api [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1449.729928] env[62235]: DEBUG oslo_vmware.api [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138285} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1449.730217] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1449.730410] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Deleted contents of the VM from datastore datastore1 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1449.730593] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1449.730768] env[62235]: INFO nova.compute.manager [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1449.731016] env[62235]: DEBUG oslo.service.loopingcall [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1449.731220] env[62235]: DEBUG nova.compute.manager [-] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1449.731314] env[62235]: DEBUG nova.network.neutron [-] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1449.952834] env[62235]: DEBUG nova.compute.manager [req-6fcbe453-8f3e-4a55-9f4c-ddcb7636982c req-0a08f18b-8774-4060-96eb-757f19fc5538 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Received event network-vif-deleted-3bfe8ae8-f413-45d7-9899-1f459beb7534 {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1449.953064] env[62235]: INFO nova.compute.manager [req-6fcbe453-8f3e-4a55-9f4c-ddcb7636982c req-0a08f18b-8774-4060-96eb-757f19fc5538 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Neutron deleted interface 3bfe8ae8-f413-45d7-9899-1f459beb7534; detaching it from the instance and deleting it from the info cache [ 1449.953329] env[62235]: DEBUG nova.network.neutron [req-6fcbe453-8f3e-4a55-9f4c-ddcb7636982c req-0a08f18b-8774-4060-96eb-757f19fc5538 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1450.432226] env[62235]: DEBUG nova.network.neutron [-] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1450.456128] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa268c06-cac5-4bf8-bd41-77f2d87187da {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.466496] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82c6c10-138b-4737-9585-54ece5b55a43 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1450.489806] env[62235]: DEBUG nova.compute.manager [req-6fcbe453-8f3e-4a55-9f4c-ddcb7636982c req-0a08f18b-8774-4060-96eb-757f19fc5538 service nova] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Detach interface failed, port_id=3bfe8ae8-f413-45d7-9899-1f459beb7534, reason: Instance e3fc3ba2-5081-455f-be58-fd8b784c4db8 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1450.728256] env[62235]: DEBUG nova.network.neutron [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updating instance_info_cache with network_info: [{"id": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "address": "fa:16:3e:73:ba:c1", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap909b7c34-70", "ovs_interfaceid": "909b7c34-704b-47db-a0c2-a33b6e25a8da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1450.935449] env[62235]: INFO nova.compute.manager [-] [instance: e3fc3ba2-5081-455f-be58-fd8b784c4db8] Took 1.20 seconds to deallocate network for instance. [ 1451.230999] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Releasing lock "refresh_cache-8e177c90-8d68-4d70-9134-d2635abdfe7f" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1451.231181] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updated the network info_cache for instance {{(pid=62235) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1451.231416] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.231599] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.231809] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.231996] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.232197] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.232379] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.232514] env[62235]: DEBUG nova.compute.manager [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62235) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1451.232699] env[62235]: DEBUG oslo_service.periodic_task [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62235) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1451.442612] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1451.442930] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1451.443201] env[62235]: DEBUG nova.objects.instance [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lazy-loading 'resources' on Instance uuid e3fc3ba2-5081-455f-be58-fd8b784c4db8 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1451.735896] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1451.988756] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d3c486-ba86-44ed-af8f-0b56615c354e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1451.996322] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4323957-f3b0-41cd-b199-558e66829dfb {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.028842] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82000f02-0270-4e13-8b48-219d0cf79d59 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.035836] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670a1e2f-02f0-4a75-9f71-28cdee1249e8 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1452.048481] env[62235]: DEBUG nova.compute.provider_tree [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1452.551511] env[62235]: DEBUG nova.scheduler.client.report [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1453.057263] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1453.059569] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.324s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1453.059768] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1453.059932] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62235) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1453.060925] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f20eb3b-dd1c-4b67-9be0-708907c8c4c0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.069133] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14631e1b-8e08-4a82-891b-6b9ca46372fa {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.074615] env[62235]: INFO nova.scheduler.client.report [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted allocations for instance e3fc3ba2-5081-455f-be58-fd8b784c4db8 [ 1453.085887] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fc1d4a-5f19-40c1-8797-c7bbcbf27372 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.092413] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655b857f-a440-4f4e-937c-96ab258adf51 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1453.122696] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181109MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62235) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1453.122833] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1453.123041] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1453.592132] env[62235]: DEBUG oslo_concurrency.lockutils [None req-86bc9706-327d-44e8-9b0f-adce9c50b6b4 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "e3fc3ba2-5081-455f-be58-fd8b784c4db8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.004s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1454.152691] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 8e177c90-8d68-4d70-9134-d2635abdfe7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62235) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1454.460839] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "20f513de-ddba-48dc-831e-3246b82213a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1454.461131] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1454.655719] env[62235]: INFO nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Instance 20f513de-ddba-48dc-831e-3246b82213a8 has allocations against this compute host but is not found in the database. [ 1454.656045] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1454.656099] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62235) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1454.691874] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19f8ace-be83-4907-ad00-305b3da3b973 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.699374] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acee036e-17dd-45ab-a63d-b1cf1cf81f8a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.729665] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f19ccb6-515a-4ae7-8774-b1ee4a4b9b34 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.736451] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e01b53b-e394-4206-8334-f033e0f49901 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1454.750090] env[62235]: DEBUG nova.compute.provider_tree [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1454.963742] env[62235]: DEBUG nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Starting instance... {{(pid=62235) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1455.253860] env[62235]: DEBUG nova.scheduler.client.report [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1455.482221] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1455.759290] env[62235]: DEBUG nova.compute.resource_tracker [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62235) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1455.759681] env[62235]: DEBUG oslo_concurrency.lockutils [None req-768b2115-67e2-48f0-82e9-fd64cc8c969d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.636s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1455.759852] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.278s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1455.761307] env[62235]: INFO nova.compute.claims [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1456.807556] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba66d9b-89bf-41d5-aea9-0af4b306f5df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1456.813431] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165a488c-e6cc-420a-bb92-405842e8974c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1456.843055] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac8b14a-3d7a-479a-b364-aff0b391bcc3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1456.849614] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4717f85d-2d91-497c-a9a2-4e9d53fe2096 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1456.862071] env[62235]: DEBUG nova.compute.provider_tree [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1457.365098] env[62235]: DEBUG nova.scheduler.client.report [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1457.869820] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1457.870394] env[62235]: DEBUG nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Start building networks asynchronously for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1458.375109] env[62235]: DEBUG nova.compute.utils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Using /dev/sd instead of None {{(pid=62235) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1458.376592] env[62235]: DEBUG nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Allocating IP information in the background. {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1458.376764] env[62235]: DEBUG nova.network.neutron [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] allocate_for_instance() {{(pid=62235) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1458.426026] env[62235]: DEBUG nova.policy [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8de8d8815704b1ca99fe1f48d746109', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd439d7dfff414de6ba781d2a7d464120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62235) authorize /opt/stack/nova/nova/policy.py:201}} [ 1458.667198] env[62235]: DEBUG nova.network.neutron [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Successfully created port: 4fd6ce51-4e62-460b-a8a9-5c87c0485b5a {{(pid=62235) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1458.880805] env[62235]: DEBUG nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Start building block device mappings for instance. {{(pid=62235) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1459.890348] env[62235]: DEBUG nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Start spawning the instance on the hypervisor. {{(pid=62235) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1459.915507] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T06:16:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T06:16:28Z,direct_url=,disk_format='vmdk',id=02c64327-1eca-4d55-9f01-62fa8b2a2334,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='52c5b3858c2a4a4c8151c52020553269',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T06:16:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1459.915762] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1459.915956] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image limits 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1459.916180] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Flavor pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1459.916339] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Image pref 0:0:0 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1459.916515] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62235) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1459.916763] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1459.916943] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1459.917132] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Got 1 possible topologies {{(pid=62235) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1459.917308] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1459.917485] env[62235]: DEBUG nova.virt.hardware [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62235) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1459.918390] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741e0358-8e71-48b1-82b3-fa8eb78a5565 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.925942] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5f94a5-7b66-4be3-9fbf-fa2650d7ee34 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.023128] env[62235]: DEBUG nova.compute.manager [req-e1c182a9-93a6-48b5-8a9b-d704732a9595 req-25c31e3b-b50a-4c7e-80ad-92d05cc9b8de service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Received event network-vif-plugged-4fd6ce51-4e62-460b-a8a9-5c87c0485b5a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1460.023380] env[62235]: DEBUG oslo_concurrency.lockutils [req-e1c182a9-93a6-48b5-8a9b-d704732a9595 req-25c31e3b-b50a-4c7e-80ad-92d05cc9b8de service nova] Acquiring lock "20f513de-ddba-48dc-831e-3246b82213a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1460.023593] env[62235]: DEBUG oslo_concurrency.lockutils [req-e1c182a9-93a6-48b5-8a9b-d704732a9595 req-25c31e3b-b50a-4c7e-80ad-92d05cc9b8de service nova] Lock "20f513de-ddba-48dc-831e-3246b82213a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1460.023844] env[62235]: DEBUG oslo_concurrency.lockutils [req-e1c182a9-93a6-48b5-8a9b-d704732a9595 req-25c31e3b-b50a-4c7e-80ad-92d05cc9b8de service nova] Lock "20f513de-ddba-48dc-831e-3246b82213a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1460.023972] env[62235]: DEBUG nova.compute.manager [req-e1c182a9-93a6-48b5-8a9b-d704732a9595 req-25c31e3b-b50a-4c7e-80ad-92d05cc9b8de service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] No waiting events found dispatching network-vif-plugged-4fd6ce51-4e62-460b-a8a9-5c87c0485b5a {{(pid=62235) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1460.024211] env[62235]: WARNING nova.compute.manager [req-e1c182a9-93a6-48b5-8a9b-d704732a9595 req-25c31e3b-b50a-4c7e-80ad-92d05cc9b8de service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Received unexpected event network-vif-plugged-4fd6ce51-4e62-460b-a8a9-5c87c0485b5a for instance with vm_state building and task_state spawning. [ 1460.106311] env[62235]: DEBUG nova.network.neutron [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Successfully updated port: 4fd6ce51-4e62-460b-a8a9-5c87c0485b5a {{(pid=62235) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1460.608863] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "refresh_cache-20f513de-ddba-48dc-831e-3246b82213a8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1460.609033] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "refresh_cache-20f513de-ddba-48dc-831e-3246b82213a8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1460.609199] env[62235]: DEBUG nova.network.neutron [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Building network info cache for instance {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1461.141170] env[62235]: DEBUG nova.network.neutron [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Instance cache missing network info. {{(pid=62235) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1461.269320] env[62235]: DEBUG nova.network.neutron [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Updating instance_info_cache with network_info: [{"id": "4fd6ce51-4e62-460b-a8a9-5c87c0485b5a", "address": "fa:16:3e:49:d1:5c", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fd6ce51-4e", "ovs_interfaceid": "4fd6ce51-4e62-460b-a8a9-5c87c0485b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1461.771616] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "refresh_cache-20f513de-ddba-48dc-831e-3246b82213a8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1461.771926] env[62235]: DEBUG nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Instance network_info: |[{"id": "4fd6ce51-4e62-460b-a8a9-5c87c0485b5a", "address": "fa:16:3e:49:d1:5c", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fd6ce51-4e", "ovs_interfaceid": "4fd6ce51-4e62-460b-a8a9-5c87c0485b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62235) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1461.772398] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:d1:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '849fc06e-dfc2-470f-8490-034590682ea7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4fd6ce51-4e62-460b-a8a9-5c87c0485b5a', 'vif_model': 'vmxnet3'}] {{(pid=62235) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1461.779853] env[62235]: DEBUG oslo.service.loopingcall [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1461.780079] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Creating VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1461.780316] env[62235]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccae0ce7-21d5-4e77-86f8-605ba48bb27f {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.799680] env[62235]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1461.799680] env[62235]: value = "task-1272420" [ 1461.799680] env[62235]: _type = "Task" [ 1461.799680] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1461.806755] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272420, 'name': CreateVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1462.049367] env[62235]: DEBUG nova.compute.manager [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Received event network-changed-4fd6ce51-4e62-460b-a8a9-5c87c0485b5a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1462.049447] env[62235]: DEBUG nova.compute.manager [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Refreshing instance network info cache due to event network-changed-4fd6ce51-4e62-460b-a8a9-5c87c0485b5a. {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1462.049671] env[62235]: DEBUG oslo_concurrency.lockutils [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] Acquiring lock "refresh_cache-20f513de-ddba-48dc-831e-3246b82213a8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1462.049818] env[62235]: DEBUG oslo_concurrency.lockutils [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] Acquired lock "refresh_cache-20f513de-ddba-48dc-831e-3246b82213a8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1462.049991] env[62235]: DEBUG nova.network.neutron [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Refreshing network info cache for port 4fd6ce51-4e62-460b-a8a9-5c87c0485b5a {{(pid=62235) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1462.310454] env[62235]: DEBUG oslo_vmware.api [-] Task: {'id': task-1272420, 'name': CreateVM_Task, 'duration_secs': 0.339351} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1462.310811] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Created VM on the ESX host {{(pid=62235) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1462.311335] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1462.311509] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1462.311835] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1462.312112] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8af549f-e17b-4297-9ad1-205062128a8b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.316749] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1462.316749] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dab06f-5e46-cd9f-13fa-c0d9c4f483fe" [ 1462.316749] env[62235]: _type = "Task" [ 1462.316749] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1462.324479] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dab06f-5e46-cd9f-13fa-c0d9c4f483fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1462.727165] env[62235]: DEBUG nova.network.neutron [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Updated VIF entry in instance network info cache for port 4fd6ce51-4e62-460b-a8a9-5c87c0485b5a. {{(pid=62235) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1462.727420] env[62235]: DEBUG nova.network.neutron [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Updating instance_info_cache with network_info: [{"id": "4fd6ce51-4e62-460b-a8a9-5c87c0485b5a", "address": "fa:16:3e:49:d1:5c", "network": {"id": "40dd6b79-0008-4ef3-8028-c9859161438a", "bridge": "br-int", "label": "tempest-ServersTestJSON-245608306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d439d7dfff414de6ba781d2a7d464120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "849fc06e-dfc2-470f-8490-034590682ea7", "external-id": "nsx-vlan-transportzone-567", "segmentation_id": 567, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fd6ce51-4e", "ovs_interfaceid": "4fd6ce51-4e62-460b-a8a9-5c87c0485b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1462.827191] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52dab06f-5e46-cd9f-13fa-c0d9c4f483fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009617} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1462.827448] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1462.827685] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Processing image 02c64327-1eca-4d55-9f01-62fa8b2a2334 {{(pid=62235) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1462.827917] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1462.828081] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1462.828277] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1462.828543] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aeca0b60-005e-4063-8975-c6102a13eaa2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.836670] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62235) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1462.836845] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62235) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1462.837529] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec84e777-03f4-497f-a34a-32c1f8a91206 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.841892] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1462.841892] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b6430b-8415-908a-1f8a-28b774ac0fb8" [ 1462.841892] env[62235]: _type = "Task" [ 1462.841892] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1462.848636] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b6430b-8415-908a-1f8a-28b774ac0fb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1463.230254] env[62235]: DEBUG oslo_concurrency.lockutils [req-44c8608c-6ec6-4e30-adc8-33f6754a11c0 req-824d3d7f-814f-45ed-8d82-00014302ad1c service nova] Releasing lock "refresh_cache-20f513de-ddba-48dc-831e-3246b82213a8" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1463.352442] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]52b6430b-8415-908a-1f8a-28b774ac0fb8, 'name': SearchDatastore_Task, 'duration_secs': 0.008938} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1463.353207] env[62235]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc61860b-dfe4-4b68-970a-08b208eca21a {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1463.357863] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1463.357863] env[62235]: value = "session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5227324c-9d22-98de-ac4a-899f4a290de7" [ 1463.357863] env[62235]: _type = "Task" [ 1463.357863] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1463.365744] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5227324c-9d22-98de-ac4a-899f4a290de7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1463.867709] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': session[52c5de13-4a13-e50c-31cb-3aeff7fca2c8]5227324c-9d22-98de-ac4a-899f4a290de7, 'name': SearchDatastore_Task, 'duration_secs': 0.008886} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1463.867923] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk" {{(pid=62235) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1463.868201] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 20f513de-ddba-48dc-831e-3246b82213a8/20f513de-ddba-48dc-831e-3246b82213a8.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1463.868478] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c81607a-9ba8-4e20-8686-46636998e2cd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1463.874655] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1463.874655] env[62235]: value = "task-1272421" [ 1463.874655] env[62235]: _type = "Task" [ 1463.874655] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1463.881929] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272421, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1464.384159] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272421, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448464} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1464.384547] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/02c64327-1eca-4d55-9f01-62fa8b2a2334/02c64327-1eca-4d55-9f01-62fa8b2a2334.vmdk to [datastore2] 20f513de-ddba-48dc-831e-3246b82213a8/20f513de-ddba-48dc-831e-3246b82213a8.vmdk {{(pid=62235) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1464.384701] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Extending root virtual disk to 1048576 {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1464.384903] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8910881a-6713-42a9-826b-a840b9b36661 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.391605] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1464.391605] env[62235]: value = "task-1272422" [ 1464.391605] env[62235]: _type = "Task" [ 1464.391605] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1464.398619] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272422, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1464.901812] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272422, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07892} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1464.902101] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Extended root virtual disk {{(pid=62235) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1464.902849] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f6fe7f-b8d1-4883-ba87-022596829d57 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.924271] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] 20f513de-ddba-48dc-831e-3246b82213a8/20f513de-ddba-48dc-831e-3246b82213a8.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1464.924521] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11da1693-ac74-43e5-bd6f-ab1e520b9cdd {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1464.943776] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1464.943776] env[62235]: value = "task-1272423" [ 1464.943776] env[62235]: _type = "Task" [ 1464.943776] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1464.950736] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272423, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1465.454670] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272423, 'name': ReconfigVM_Task, 'duration_secs': 0.286849} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1465.455902] env[62235]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Reconfigured VM instance instance-00000074 to attach disk [datastore2] 20f513de-ddba-48dc-831e-3246b82213a8/20f513de-ddba-48dc-831e-3246b82213a8.vmdk or device None with type sparse {{(pid=62235) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1465.457019] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d0783dc-92d9-4b27-a153-60aca2386821 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.463191] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1465.463191] env[62235]: value = "task-1272424" [ 1465.463191] env[62235]: _type = "Task" [ 1465.463191] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1465.470376] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272424, 'name': Rename_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1465.974629] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272424, 'name': Rename_Task, 'duration_secs': 0.146251} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1465.974933] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Powering on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1465.975203] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1ae78f8-f90e-4650-8f8b-4997392f5177 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.980955] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1465.980955] env[62235]: value = "task-1272425" [ 1465.980955] env[62235]: _type = "Task" [ 1465.980955] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1465.988250] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272425, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1466.491585] env[62235]: DEBUG oslo_vmware.api [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272425, 'name': PowerOnVM_Task, 'duration_secs': 0.418412} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1466.492022] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Powered on the VM {{(pid=62235) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1466.492078] env[62235]: INFO nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Took 6.60 seconds to spawn the instance on the hypervisor. [ 1466.492262] env[62235]: DEBUG nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1466.493025] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e6b972-767e-407d-918d-0882d56e7589 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.010935] env[62235]: INFO nova.compute.manager [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Took 11.54 seconds to build instance. [ 1467.512862] env[62235]: DEBUG oslo_concurrency.lockutils [None req-cd15738c-68a1-4272-a423-a97d33d85bb6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.052s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1467.879173] env[62235]: DEBUG oslo_concurrency.lockutils [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "20f513de-ddba-48dc-831e-3246b82213a8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1467.879440] env[62235]: DEBUG oslo_concurrency.lockutils [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1467.879674] env[62235]: DEBUG nova.compute.manager [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1467.880575] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c494a2a8-3b71-434f-b6dd-6fb389f5fff9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.888112] env[62235]: DEBUG nova.compute.manager [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62235) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1467.888258] env[62235]: DEBUG nova.objects.instance [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lazy-loading 'flavor' on Instance uuid 20f513de-ddba-48dc-831e-3246b82213a8 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1468.393823] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1468.394128] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44c939fb-5ee1-49e8-9fdb-025415d9c7df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.401100] env[62235]: DEBUG oslo_vmware.api [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1468.401100] env[62235]: value = "task-1272426" [ 1468.401100] env[62235]: _type = "Task" [ 1468.401100] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1468.409084] env[62235]: DEBUG oslo_vmware.api [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272426, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1468.913022] env[62235]: DEBUG oslo_vmware.api [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272426, 'name': PowerOffVM_Task, 'duration_secs': 0.205335} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1468.913406] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1468.913541] env[62235]: DEBUG nova.compute.manager [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Checking state {{(pid=62235) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1468.914303] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47b1e0c-9fc7-4da6-a958-f5c6e460bcf3 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.425651] env[62235]: DEBUG oslo_concurrency.lockutils [None req-00c685e1-e3e8-4e36-a921-2456149e61f6 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1470.506689] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "20f513de-ddba-48dc-831e-3246b82213a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1470.507115] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1470.507216] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "20f513de-ddba-48dc-831e-3246b82213a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1470.507426] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1470.507609] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1470.509836] env[62235]: INFO nova.compute.manager [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Terminating instance [ 1470.511627] env[62235]: DEBUG nova.compute.manager [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1470.511823] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1470.512654] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3135f762-931c-461e-8a10-b2c841761d80 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.519954] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1470.520448] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f769720b-eae5-4441-b764-5b13515dac19 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.581193] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1470.581425] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1470.581615] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleting the datastore file [datastore2] 20f513de-ddba-48dc-831e-3246b82213a8 {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1470.581873] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cae1d0c4-6703-4cdb-b34c-95c4c47fcf58 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.588387] env[62235]: DEBUG oslo_vmware.api [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1470.588387] env[62235]: value = "task-1272428" [ 1470.588387] env[62235]: _type = "Task" [ 1470.588387] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1470.595643] env[62235]: DEBUG oslo_vmware.api [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1471.104905] env[62235]: DEBUG oslo_vmware.api [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133696} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1471.105302] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1471.105572] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1471.105816] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1471.106066] env[62235]: INFO nova.compute.manager [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1471.106404] env[62235]: DEBUG oslo.service.loopingcall [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1471.106654] env[62235]: DEBUG nova.compute.manager [-] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1471.106779] env[62235]: DEBUG nova.network.neutron [-] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1471.441319] env[62235]: DEBUG nova.compute.manager [req-d68c13a8-ea8c-460f-986f-32b9f73b256f req-584b3ece-4551-40c4-9a6e-b3e40d995684 service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Received event network-vif-deleted-4fd6ce51-4e62-460b-a8a9-5c87c0485b5a {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1471.441319] env[62235]: INFO nova.compute.manager [req-d68c13a8-ea8c-460f-986f-32b9f73b256f req-584b3ece-4551-40c4-9a6e-b3e40d995684 service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Neutron deleted interface 4fd6ce51-4e62-460b-a8a9-5c87c0485b5a; detaching it from the instance and deleting it from the info cache [ 1471.441319] env[62235]: DEBUG nova.network.neutron [req-d68c13a8-ea8c-460f-986f-32b9f73b256f req-584b3ece-4551-40c4-9a6e-b3e40d995684 service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1471.918842] env[62235]: DEBUG nova.network.neutron [-] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1471.943298] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21696170-334e-4e4f-9608-dcf664b6bf46 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1471.953103] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7b2e01-8b34-43d8-86a4-c8b4752380c2 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1471.975489] env[62235]: DEBUG nova.compute.manager [req-d68c13a8-ea8c-460f-986f-32b9f73b256f req-584b3ece-4551-40c4-9a6e-b3e40d995684 service nova] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Detach interface failed, port_id=4fd6ce51-4e62-460b-a8a9-5c87c0485b5a, reason: Instance 20f513de-ddba-48dc-831e-3246b82213a8 could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1472.421298] env[62235]: INFO nova.compute.manager [-] [instance: 20f513de-ddba-48dc-831e-3246b82213a8] Took 1.31 seconds to deallocate network for instance. [ 1472.928165] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1472.928474] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1472.928686] env[62235]: DEBUG nova.objects.instance [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lazy-loading 'resources' on Instance uuid 20f513de-ddba-48dc-831e-3246b82213a8 {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1473.471591] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d64e4a-fa13-400c-a843-de8bbb4be37b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.479017] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38df9d02-4b80-4e7b-bb19-f27e69706c6e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.509120] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a7a594-db93-4ed0-bd58-6ed2ee09b67c {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.515545] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf5d093-095c-4013-85df-b23edf02d2e9 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.528224] env[62235]: DEBUG nova.compute.provider_tree [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1474.031511] env[62235]: DEBUG nova.scheduler.client.report [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1474.537231] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1474.559183] env[62235]: INFO nova.scheduler.client.report [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted allocations for instance 20f513de-ddba-48dc-831e-3246b82213a8 [ 1475.068338] env[62235]: DEBUG oslo_concurrency.lockutils [None req-28c88468-ecd2-4392-b896-0c73880f99d9 tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "20f513de-ddba-48dc-831e-3246b82213a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.561s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1476.284057] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8e177c90-8d68-4d70-9134-d2635abdfe7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1476.284057] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1476.284484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "8e177c90-8d68-4d70-9134-d2635abdfe7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1476.284484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1476.284484] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1476.286670] env[62235]: INFO nova.compute.manager [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Terminating instance [ 1476.288439] env[62235]: DEBUG nova.compute.manager [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Start destroying the instance on the hypervisor. {{(pid=62235) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1476.288650] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Destroying instance {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1476.289502] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb1fee7-feae-49c3-a156-97aa58a715a0 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.297496] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Powering off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1476.298015] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-766a3983-01be-4238-89ea-507845efb4d1 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.303852] env[62235]: DEBUG oslo_vmware.api [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1476.303852] env[62235]: value = "task-1272429" [ 1476.303852] env[62235]: _type = "Task" [ 1476.303852] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1476.311894] env[62235]: DEBUG oslo_vmware.api [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1476.813802] env[62235]: DEBUG oslo_vmware.api [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272429, 'name': PowerOffVM_Task, 'duration_secs': 0.204225} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1476.814078] env[62235]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Powered off the VM {{(pid=62235) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1476.814258] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Unregistering the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1476.814500] env[62235]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f049f96-47d5-4a10-a1f9-38057bc19229 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.876541] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Unregistered the VM {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1476.876706] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Deleting contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1476.876918] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleting the datastore file [datastore2] 8e177c90-8d68-4d70-9134-d2635abdfe7f {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1476.877169] env[62235]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-138d68aa-5b3f-4f76-a86a-a37f3ac89c0e {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.883813] env[62235]: DEBUG oslo_vmware.api [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for the task: (returnval){ [ 1476.883813] env[62235]: value = "task-1272431" [ 1476.883813] env[62235]: _type = "Task" [ 1476.883813] env[62235]: } to complete. {{(pid=62235) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1476.890797] env[62235]: DEBUG oslo_vmware.api [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1477.394662] env[62235]: DEBUG oslo_vmware.api [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Task: {'id': task-1272431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135917} completed successfully. {{(pid=62235) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1477.396040] env[62235]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted the datastore file {{(pid=62235) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1477.396040] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Deleted contents of the VM from datastore datastore2 {{(pid=62235) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1477.396040] env[62235]: DEBUG nova.virt.vmwareapi.vmops [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Instance destroyed {{(pid=62235) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1477.396040] env[62235]: INFO nova.compute.manager [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1477.396327] env[62235]: DEBUG oslo.service.loopingcall [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62235) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1477.396512] env[62235]: DEBUG nova.compute.manager [-] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Deallocating network for instance {{(pid=62235) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1477.396612] env[62235]: DEBUG nova.network.neutron [-] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] deallocate_for_instance() {{(pid=62235) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1477.637658] env[62235]: DEBUG nova.compute.manager [req-ebe78091-accb-4c54-aebe-56c8616f84bb req-f9aa4dd2-f360-47ef-9496-76177663a376 service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Received event network-vif-deleted-909b7c34-704b-47db-a0c2-a33b6e25a8da {{(pid=62235) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1477.637832] env[62235]: INFO nova.compute.manager [req-ebe78091-accb-4c54-aebe-56c8616f84bb req-f9aa4dd2-f360-47ef-9496-76177663a376 service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Neutron deleted interface 909b7c34-704b-47db-a0c2-a33b6e25a8da; detaching it from the instance and deleting it from the info cache [ 1477.637971] env[62235]: DEBUG nova.network.neutron [req-ebe78091-accb-4c54-aebe-56c8616f84bb req-f9aa4dd2-f360-47ef-9496-76177663a376 service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1478.111465] env[62235]: DEBUG nova.network.neutron [-] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Updating instance_info_cache with network_info: [] {{(pid=62235) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1478.141840] env[62235]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5f82bc8-9b59-4cdb-8991-d3107e17894d {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1478.151417] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3955e1-e2be-48c8-b29e-bfe953d789df {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1478.172968] env[62235]: DEBUG nova.compute.manager [req-ebe78091-accb-4c54-aebe-56c8616f84bb req-f9aa4dd2-f360-47ef-9496-76177663a376 service nova] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Detach interface failed, port_id=909b7c34-704b-47db-a0c2-a33b6e25a8da, reason: Instance 8e177c90-8d68-4d70-9134-d2635abdfe7f could not be found. {{(pid=62235) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1478.614273] env[62235]: INFO nova.compute.manager [-] [instance: 8e177c90-8d68-4d70-9134-d2635abdfe7f] Took 1.22 seconds to deallocate network for instance. [ 1479.120881] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1479.121177] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1479.121407] env[62235]: DEBUG nova.objects.instance [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lazy-loading 'resources' on Instance uuid 8e177c90-8d68-4d70-9134-d2635abdfe7f {{(pid=62235) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1479.662198] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b136434c-6495-4ca0-9dbd-231bb6ea2129 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.670154] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecea3a4-cbc4-4524-bbb9-d01c6f42b2ac {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.701102] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b10ee1-4220-4788-9c20-bbfb3a574d73 {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.708806] env[62235]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2260e16e-0347-41e9-84c2-1dbb04055d2b {{(pid=62235) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.724153] env[62235]: DEBUG nova.compute.provider_tree [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed in ProviderTree for provider: 108a4390-ff68-4048-b61d-c7a9614ddc4f {{(pid=62235) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1480.227609] env[62235]: DEBUG nova.scheduler.client.report [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Inventory has not changed for provider 108a4390-ff68-4048-b61d-c7a9614ddc4f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62235) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1480.732028] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1480.750880] env[62235]: INFO nova.scheduler.client.report [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Deleted allocations for instance 8e177c90-8d68-4d70-9134-d2635abdfe7f [ 1481.258312] env[62235]: DEBUG oslo_concurrency.lockutils [None req-ed517f35-0fdb-43e8-9976-11e2d773be1c tempest-ServersTestJSON-1672881899 tempest-ServersTestJSON-1672881899-project-member] Lock "8e177c90-8d68-4d70-9134-d2635abdfe7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.974s {{(pid=62235) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}